Oct 08 00:06:40 crc systemd[1]: Starting Kubernetes Kubelet... Oct 08 00:06:40 crc restorecon[4670]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:40 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 00:06:41 crc restorecon[4670]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 00:06:41 crc restorecon[4670]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Oct 08 00:06:41 crc kubenswrapper[4967]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 08 00:06:41 crc kubenswrapper[4967]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Oct 08 00:06:41 crc kubenswrapper[4967]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 08 00:06:41 crc kubenswrapper[4967]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 08 00:06:41 crc kubenswrapper[4967]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Oct 08 00:06:41 crc kubenswrapper[4967]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.842366 4967 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.848262 4967 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.848298 4967 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.848308 4967 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.848317 4967 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.848327 4967 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.848336 4967 feature_gate.go:330] unrecognized feature gate: Example Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.848344 4967 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.848353 4967 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.848362 4967 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.848370 4967 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.848378 4967 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.848385 4967 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.848394 4967 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.848402 4967 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.848410 4967 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.848420 4967 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.848440 4967 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.848449 4967 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.848459 4967 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.848468 4967 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.848477 4967 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.848486 4967 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.848494 4967 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.848502 4967 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.848509 4967 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.848521 4967 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.848531 4967 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.848540 4967 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.848548 4967 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.848557 4967 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.848565 4967 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.848573 4967 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.848581 4967 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.848590 4967 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.848598 4967 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.848605 4967 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.848613 4967 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.848621 4967 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.848630 4967 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.848638 4967 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.848646 4967 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.848654 4967 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.848662 4967 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.848697 4967 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.848705 4967 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.848713 4967 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.848720 4967 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.848730 4967 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.848738 4967 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.848746 4967 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.848753 4967 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.848763 4967 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.848775 4967 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.848785 4967 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.848795 4967 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.848803 4967 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.848813 4967 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.848823 4967 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.848831 4967 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.848842 4967 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.848851 4967 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.848860 4967 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.848868 4967 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.848877 4967 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.848886 4967 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.848894 4967 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.848902 4967 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.848909 4967 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.848917 4967 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.848925 4967 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.848933 4967 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850026 4967 flags.go:64] FLAG: --address="0.0.0.0" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850054 4967 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850071 4967 flags.go:64] FLAG: --anonymous-auth="true" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850083 4967 flags.go:64] FLAG: --application-metrics-count-limit="100" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850095 4967 flags.go:64] FLAG: --authentication-token-webhook="false" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850104 4967 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850116 4967 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850127 4967 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850137 4967 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850146 4967 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850156 4967 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850166 4967 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850175 4967 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850184 4967 flags.go:64] FLAG: --cgroup-root="" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850193 4967 flags.go:64] FLAG: --cgroups-per-qos="true" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850202 4967 flags.go:64] FLAG: --client-ca-file="" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850211 4967 flags.go:64] FLAG: --cloud-config="" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850221 4967 flags.go:64] FLAG: --cloud-provider="" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850230 4967 flags.go:64] FLAG: --cluster-dns="[]" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850241 4967 flags.go:64] FLAG: --cluster-domain="" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850250 4967 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850260 4967 flags.go:64] FLAG: --config-dir="" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850268 4967 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850278 4967 flags.go:64] FLAG: --container-log-max-files="5" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850290 4967 flags.go:64] FLAG: --container-log-max-size="10Mi" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850299 4967 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850308 4967 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850317 4967 flags.go:64] FLAG: --containerd-namespace="k8s.io" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850327 4967 flags.go:64] FLAG: --contention-profiling="false" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850336 4967 flags.go:64] FLAG: --cpu-cfs-quota="true" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850345 4967 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850356 4967 flags.go:64] FLAG: --cpu-manager-policy="none" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850367 4967 flags.go:64] FLAG: --cpu-manager-policy-options="" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850380 4967 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850391 4967 flags.go:64] FLAG: --enable-controller-attach-detach="true" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850401 4967 flags.go:64] FLAG: --enable-debugging-handlers="true" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850412 4967 flags.go:64] FLAG: --enable-load-reader="false" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850421 4967 flags.go:64] FLAG: --enable-server="true" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850429 4967 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850442 4967 flags.go:64] FLAG: --event-burst="100" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850452 4967 flags.go:64] FLAG: --event-qps="50" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850461 4967 flags.go:64] FLAG: --event-storage-age-limit="default=0" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850470 4967 flags.go:64] FLAG: --event-storage-event-limit="default=0" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850480 4967 flags.go:64] FLAG: --eviction-hard="" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850491 4967 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850500 4967 flags.go:64] FLAG: --eviction-minimum-reclaim="" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850509 4967 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850518 4967 flags.go:64] FLAG: --eviction-soft="" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850526 4967 flags.go:64] FLAG: --eviction-soft-grace-period="" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850535 4967 flags.go:64] FLAG: --exit-on-lock-contention="false" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850544 4967 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850553 4967 flags.go:64] FLAG: --experimental-mounter-path="" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850561 4967 flags.go:64] FLAG: --fail-cgroupv1="false" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850572 4967 flags.go:64] FLAG: --fail-swap-on="true" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850581 4967 flags.go:64] FLAG: --feature-gates="" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850592 4967 flags.go:64] FLAG: --file-check-frequency="20s" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850601 4967 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850610 4967 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850621 4967 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850630 4967 flags.go:64] FLAG: --healthz-port="10248" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850639 4967 flags.go:64] FLAG: --help="false" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850648 4967 flags.go:64] FLAG: --hostname-override="" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850656 4967 flags.go:64] FLAG: --housekeeping-interval="10s" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850691 4967 flags.go:64] FLAG: --http-check-frequency="20s" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850701 4967 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850709 4967 flags.go:64] FLAG: --image-credential-provider-config="" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850718 4967 flags.go:64] FLAG: --image-gc-high-threshold="85" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850727 4967 flags.go:64] FLAG: --image-gc-low-threshold="80" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850736 4967 flags.go:64] FLAG: --image-service-endpoint="" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850744 4967 flags.go:64] FLAG: --kernel-memcg-notification="false" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850753 4967 flags.go:64] FLAG: --kube-api-burst="100" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850762 4967 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850771 4967 flags.go:64] FLAG: --kube-api-qps="50" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850781 4967 flags.go:64] FLAG: --kube-reserved="" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850790 4967 flags.go:64] FLAG: --kube-reserved-cgroup="" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850799 4967 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850808 4967 flags.go:64] FLAG: --kubelet-cgroups="" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850817 4967 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850826 4967 flags.go:64] FLAG: --lock-file="" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850834 4967 flags.go:64] FLAG: --log-cadvisor-usage="false" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850843 4967 flags.go:64] FLAG: --log-flush-frequency="5s" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850852 4967 flags.go:64] FLAG: --log-json-info-buffer-size="0" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850866 4967 flags.go:64] FLAG: --log-json-split-stream="false" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850874 4967 flags.go:64] FLAG: --log-text-info-buffer-size="0" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850883 4967 flags.go:64] FLAG: --log-text-split-stream="false" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850892 4967 flags.go:64] FLAG: --logging-format="text" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850903 4967 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850912 4967 flags.go:64] FLAG: --make-iptables-util-chains="true" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850921 4967 flags.go:64] FLAG: --manifest-url="" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850932 4967 flags.go:64] FLAG: --manifest-url-header="" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850944 4967 flags.go:64] FLAG: --max-housekeeping-interval="15s" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850953 4967 flags.go:64] FLAG: --max-open-files="1000000" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850964 4967 flags.go:64] FLAG: --max-pods="110" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850973 4967 flags.go:64] FLAG: --maximum-dead-containers="-1" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850982 4967 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.850991 4967 flags.go:64] FLAG: --memory-manager-policy="None" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.851001 4967 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.851011 4967 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.851021 4967 flags.go:64] FLAG: --node-ip="192.168.126.11" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.851031 4967 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.851052 4967 flags.go:64] FLAG: --node-status-max-images="50" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.851060 4967 flags.go:64] FLAG: --node-status-update-frequency="10s" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.851070 4967 flags.go:64] FLAG: --oom-score-adj="-999" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.851079 4967 flags.go:64] FLAG: --pod-cidr="" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.851087 4967 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.851101 4967 flags.go:64] FLAG: --pod-manifest-path="" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.851109 4967 flags.go:64] FLAG: --pod-max-pids="-1" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.851119 4967 flags.go:64] FLAG: --pods-per-core="0" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.851128 4967 flags.go:64] FLAG: --port="10250" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.851137 4967 flags.go:64] FLAG: --protect-kernel-defaults="false" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.851146 4967 flags.go:64] FLAG: --provider-id="" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.851154 4967 flags.go:64] FLAG: --qos-reserved="" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.851163 4967 flags.go:64] FLAG: --read-only-port="10255" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.851172 4967 flags.go:64] FLAG: --register-node="true" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.851181 4967 flags.go:64] FLAG: --register-schedulable="true" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.851190 4967 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.851205 4967 flags.go:64] FLAG: --registry-burst="10" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.851213 4967 flags.go:64] FLAG: --registry-qps="5" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.851223 4967 flags.go:64] FLAG: --reserved-cpus="" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.851232 4967 flags.go:64] FLAG: --reserved-memory="" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.851243 4967 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.851252 4967 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.851262 4967 flags.go:64] FLAG: --rotate-certificates="false" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.851271 4967 flags.go:64] FLAG: --rotate-server-certificates="false" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.851280 4967 flags.go:64] FLAG: --runonce="false" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.851290 4967 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.851299 4967 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.851308 4967 flags.go:64] FLAG: --seccomp-default="false" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.851317 4967 flags.go:64] FLAG: --serialize-image-pulls="true" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.851327 4967 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.851336 4967 flags.go:64] FLAG: --storage-driver-db="cadvisor" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.851345 4967 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.851354 4967 flags.go:64] FLAG: --storage-driver-password="root" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.851362 4967 flags.go:64] FLAG: --storage-driver-secure="false" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.851372 4967 flags.go:64] FLAG: --storage-driver-table="stats" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.851381 4967 flags.go:64] FLAG: --storage-driver-user="root" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.851390 4967 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.851399 4967 flags.go:64] FLAG: --sync-frequency="1m0s" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.851408 4967 flags.go:64] FLAG: --system-cgroups="" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.851417 4967 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.851430 4967 flags.go:64] FLAG: --system-reserved-cgroup="" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.851439 4967 flags.go:64] FLAG: --tls-cert-file="" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.851447 4967 flags.go:64] FLAG: --tls-cipher-suites="[]" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.851458 4967 flags.go:64] FLAG: --tls-min-version="" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.851467 4967 flags.go:64] FLAG: --tls-private-key-file="" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.851476 4967 flags.go:64] FLAG: --topology-manager-policy="none" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.851484 4967 flags.go:64] FLAG: --topology-manager-policy-options="" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.851493 4967 flags.go:64] FLAG: --topology-manager-scope="container" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.851502 4967 flags.go:64] FLAG: --v="2" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.851513 4967 flags.go:64] FLAG: --version="false" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.851525 4967 flags.go:64] FLAG: --vmodule="" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.851535 4967 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.851545 4967 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.851802 4967 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.851817 4967 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.851829 4967 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.851837 4967 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.851846 4967 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.851854 4967 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.851862 4967 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.851870 4967 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.851880 4967 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.851888 4967 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.851896 4967 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.851904 4967 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.851912 4967 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.851919 4967 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.851927 4967 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.851935 4967 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.851943 4967 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.851951 4967 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.851958 4967 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.851966 4967 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.851974 4967 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.851982 4967 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.851989 4967 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.851997 4967 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.852007 4967 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.852016 4967 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.852026 4967 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.852035 4967 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.852043 4967 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.852051 4967 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.852059 4967 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.852067 4967 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.852075 4967 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.852084 4967 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.852092 4967 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.852100 4967 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.852107 4967 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.852115 4967 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.852123 4967 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.852130 4967 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.852138 4967 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.852148 4967 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.852155 4967 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.852163 4967 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.852172 4967 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.852179 4967 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.852187 4967 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.852194 4967 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.852202 4967 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.852210 4967 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.852218 4967 feature_gate.go:330] unrecognized feature gate: Example Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.852226 4967 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.852234 4967 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.852242 4967 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.852251 4967 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.852260 4967 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.852268 4967 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.852276 4967 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.852284 4967 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.852292 4967 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.852300 4967 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.852307 4967 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.852315 4967 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.852322 4967 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.852332 4967 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.852342 4967 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.852350 4967 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.852360 4967 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.852370 4967 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.852378 4967 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.852387 4967 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.852399 4967 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.867357 4967 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.867424 4967 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.867579 4967 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.867595 4967 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.867605 4967 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.867617 4967 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.867626 4967 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.867637 4967 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.867646 4967 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.867658 4967 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.867708 4967 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.867719 4967 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.867728 4967 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.867738 4967 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.867747 4967 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.867757 4967 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.867769 4967 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.867780 4967 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.867792 4967 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.867803 4967 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.867814 4967 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.867824 4967 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.867834 4967 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.867843 4967 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.867852 4967 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.867862 4967 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.867872 4967 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.867883 4967 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.867896 4967 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.867908 4967 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.867919 4967 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.867931 4967 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.867940 4967 feature_gate.go:330] unrecognized feature gate: Example Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.867952 4967 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.867965 4967 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.867974 4967 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.867984 4967 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.867993 4967 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.868003 4967 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.868011 4967 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.868019 4967 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.868028 4967 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.868037 4967 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.868045 4967 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.868054 4967 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.868063 4967 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.868071 4967 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.868079 4967 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.868087 4967 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.868096 4967 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.868105 4967 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.868115 4967 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.868125 4967 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.868134 4967 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.868143 4967 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.868151 4967 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.868159 4967 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.868171 4967 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.868182 4967 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.868192 4967 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.868201 4967 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.868210 4967 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.868219 4967 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.868228 4967 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.868236 4967 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.868247 4967 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.868256 4967 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.868265 4967 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.868275 4967 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.868284 4967 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.868292 4967 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.868300 4967 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.868309 4967 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.868323 4967 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.868570 4967 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.868586 4967 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.868597 4967 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.868606 4967 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.868615 4967 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.868623 4967 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.868632 4967 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.868641 4967 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.868650 4967 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.868661 4967 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.868701 4967 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.868710 4967 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.868722 4967 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.868732 4967 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.868742 4967 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.868751 4967 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.868759 4967 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.868770 4967 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.868779 4967 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.868788 4967 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.868797 4967 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.868806 4967 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.868814 4967 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.868823 4967 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.868833 4967 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.868842 4967 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.868850 4967 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.868859 4967 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.868868 4967 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.868879 4967 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.868890 4967 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.868901 4967 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.868911 4967 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.868925 4967 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.868937 4967 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.868946 4967 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.868955 4967 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.868968 4967 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.868978 4967 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.868986 4967 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.868994 4967 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.869003 4967 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.869011 4967 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.869019 4967 feature_gate.go:330] unrecognized feature gate: Example Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.869027 4967 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.869072 4967 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.869083 4967 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.869091 4967 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.869100 4967 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.869110 4967 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.869121 4967 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.869129 4967 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.869138 4967 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.869146 4967 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.869155 4967 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.869163 4967 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.869172 4967 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.869181 4967 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.869189 4967 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.869198 4967 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.869206 4967 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.869217 4967 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.869225 4967 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.869233 4967 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.869241 4967 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.869250 4967 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.869258 4967 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.869267 4967 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.869278 4967 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.869288 4967 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 08 00:06:41 crc kubenswrapper[4967]: W1008 00:06:41.869298 4967 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.869313 4967 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.871607 4967 server.go:940] "Client rotation is on, will bootstrap in background" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.878118 4967 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.878318 4967 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.880894 4967 server.go:997] "Starting client certificate rotation" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.880949 4967 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.882260 4967 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-27 04:48:18.576313815 +0000 UTC Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.882407 4967 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 1924h41m36.693912006s for next certificate rotation Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.911583 4967 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.914842 4967 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.933859 4967 log.go:25] "Validated CRI v1 runtime API" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.973531 4967 log.go:25] "Validated CRI v1 image API" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.975794 4967 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.984412 4967 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-10-08-00-00-42-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Oct 08 00:06:41 crc kubenswrapper[4967]: I1008 00:06:41.984465 4967 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.014758 4967 manager.go:217] Machine: {Timestamp:2025-10-08 00:06:42.011148128 +0000 UTC m=+0.661682847 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2799998 MemoryCapacity:33654124544 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:b243ca1c-20cd-4e6f-96b2-52fce74dd2d4 BootID:5cddad4c-22d1-42d7-a373-87b47c1cc97b Filesystems:[{Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365408768 Type:vfs Inodes:821633 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108169 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827060224 Type:vfs Inodes:4108169 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:af:9e:40 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:af:9e:40 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:93:7d:71 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:e0:12:4d Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:86:fc:16 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:99:f5:a7 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:b6:04:89:57:fd:7e Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:1a:1c:aa:cc:a3:b4 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654124544 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.015162 4967 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.015417 4967 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.018969 4967 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.019343 4967 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.019391 4967 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.019797 4967 topology_manager.go:138] "Creating topology manager with none policy" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.019817 4967 container_manager_linux.go:303] "Creating device plugin manager" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.020571 4967 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.020623 4967 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.021116 4967 state_mem.go:36] "Initialized new in-memory state store" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.021642 4967 server.go:1245] "Using root directory" path="/var/lib/kubelet" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.025702 4967 kubelet.go:418] "Attempting to sync node with API server" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.025748 4967 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.025778 4967 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.025803 4967 kubelet.go:324] "Adding apiserver pod source" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.025822 4967 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Oct 08 00:06:42 crc kubenswrapper[4967]: W1008 00:06:42.033159 4967 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.20:6443: connect: connection refused Oct 08 00:06:42 crc kubenswrapper[4967]: W1008 00:06:42.033222 4967 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.20:6443: connect: connection refused Oct 08 00:06:42 crc kubenswrapper[4967]: E1008 00:06:42.033304 4967 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.20:6443: connect: connection refused" logger="UnhandledError" Oct 08 00:06:42 crc kubenswrapper[4967]: E1008 00:06:42.033348 4967 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.20:6443: connect: connection refused" logger="UnhandledError" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.034347 4967 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.035614 4967 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.038242 4967 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.039849 4967 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.039878 4967 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.039886 4967 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.039893 4967 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.039906 4967 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.039915 4967 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.039922 4967 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.039933 4967 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.039943 4967 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.039951 4967 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.039965 4967 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.039972 4967 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.039994 4967 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.040592 4967 server.go:1280] "Started kubelet" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.040884 4967 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.040875 4967 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.042533 4967 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.042737 4967 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.20:6443: connect: connection refused Oct 08 00:06:42 crc systemd[1]: Started Kubernetes Kubelet. Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.044187 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.044248 4967 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.044558 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-18 13:53:07.243598067 +0000 UTC Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.044695 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 2461h46m25.198935572s for next certificate rotation Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.044749 4967 volume_manager.go:287] "The desired_state_of_world populator starts" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.044788 4967 volume_manager.go:289] "Starting Kubelet Volume Manager" Oct 08 00:06:42 crc kubenswrapper[4967]: E1008 00:06:42.044776 4967 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.044888 4967 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.045247 4967 server.go:460] "Adding debug handlers to kubelet server" Oct 08 00:06:42 crc kubenswrapper[4967]: E1008 00:06:42.046221 4967 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.20:6443: connect: connection refused" interval="200ms" Oct 08 00:06:42 crc kubenswrapper[4967]: W1008 00:06:42.046628 4967 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.20:6443: connect: connection refused Oct 08 00:06:42 crc kubenswrapper[4967]: E1008 00:06:42.046778 4967 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.20:6443: connect: connection refused" logger="UnhandledError" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.049816 4967 factory.go:55] Registering systemd factory Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.049998 4967 factory.go:221] Registration of the systemd container factory successfully Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.054320 4967 factory.go:153] Registering CRI-O factory Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.054373 4967 factory.go:221] Registration of the crio container factory successfully Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.054607 4967 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.054661 4967 factory.go:103] Registering Raw factory Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.054735 4967 manager.go:1196] Started watching for new ooms in manager Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.059945 4967 manager.go:319] Starting recovery of all containers Oct 08 00:06:42 crc kubenswrapper[4967]: E1008 00:06:42.061905 4967 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.20:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.186c5b5b4a3b1af8 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-10-08 00:06:42.040560376 +0000 UTC m=+0.691095035,LastTimestamp:2025-10-08 00:06:42.040560376 +0000 UTC m=+0.691095035,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.065264 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.065394 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.065421 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.065443 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.065467 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.065498 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.065516 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.065535 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.065559 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.065580 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.067347 4967 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.067448 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.067481 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.067542 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.067579 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.067609 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.067757 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.067781 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.067833 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.067883 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.067937 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.067985 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.068008 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.068052 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.068073 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.068092 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.068114 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.068169 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.068217 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.068237 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.068283 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.068327 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.068348 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.068394 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.068476 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.068554 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.068575 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.068650 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.068694 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.068741 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.068763 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.068809 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.068918 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.068937 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.068982 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.069029 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.069049 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.069068 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.069090 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.069166 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.069215 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.069234 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.069279 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.069340 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.069390 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.069434 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.069456 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.069500 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.069518 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.069536 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.069556 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.069575 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.069658 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.069718 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.069783 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.069805 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.069824 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.069846 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.069866 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.069885 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.069905 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.069926 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.069951 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.069972 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.069993 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.070014 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.070035 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.070054 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.070074 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.070094 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.070115 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.070136 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.070157 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.070177 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.070195 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.070215 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.070233 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.070253 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.070276 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.070295 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.070315 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.070335 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.070355 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.070374 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.070394 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.070414 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.070435 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.070454 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.070474 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.070492 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.070514 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.070533 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.070551 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.070576 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.070596 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.070620 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.070642 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.070664 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.070712 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.070732 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.070753 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.070773 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.070793 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.070814 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.070833 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.070852 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.070870 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.070888 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.070907 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.070925 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.070945 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.070964 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.070982 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.071001 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.071020 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.071039 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.071057 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.071074 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.071157 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.071176 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.071194 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.071219 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.071238 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.071255 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.071273 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.071321 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.071340 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.071357 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.071374 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.071393 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.071466 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.071486 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.071505 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.071523 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.071542 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.071561 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.071580 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.071598 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.071615 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.071634 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.071653 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.071699 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.071723 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.071742 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.071760 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.071778 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.071796 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.071813 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.071830 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.071847 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.071868 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.071886 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.071904 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.071921 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.071938 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.071955 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.071973 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.071991 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.072011 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.072030 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.072048 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.072083 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.072103 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.072123 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.072142 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.072161 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.072182 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.072199 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.072217 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.072236 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.072255 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.072271 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.072289 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.072313 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.072336 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.072355 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.072372 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.072391 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.072410 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.072428 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.072449 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.072466 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.072486 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.072503 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.072523 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.072542 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.072561 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.072578 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.072597 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.072615 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.072635 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.072652 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.072735 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.072754 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.072774 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.072794 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.072812 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.072829 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.072849 4967 reconstruct.go:97] "Volume reconstruction finished" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.072862 4967 reconciler.go:26] "Reconciler: start to sync state" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.089000 4967 manager.go:324] Recovery completed Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.106322 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.108482 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.109364 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.109469 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.110830 4967 cpu_manager.go:225] "Starting CPU manager" policy="none" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.110851 4967 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.110874 4967 state_mem.go:36] "Initialized new in-memory state store" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.113806 4967 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.116485 4967 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.116573 4967 status_manager.go:217] "Starting to sync pod status with apiserver" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.116624 4967 kubelet.go:2335] "Starting kubelet main sync loop" Oct 08 00:06:42 crc kubenswrapper[4967]: E1008 00:06:42.116772 4967 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Oct 08 00:06:42 crc kubenswrapper[4967]: W1008 00:06:42.118613 4967 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.20:6443: connect: connection refused Oct 08 00:06:42 crc kubenswrapper[4967]: E1008 00:06:42.118713 4967 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.20:6443: connect: connection refused" logger="UnhandledError" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.132149 4967 policy_none.go:49] "None policy: Start" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.133456 4967 memory_manager.go:170] "Starting memorymanager" policy="None" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.133487 4967 state_mem.go:35] "Initializing new in-memory state store" Oct 08 00:06:42 crc kubenswrapper[4967]: E1008 00:06:42.145412 4967 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.199318 4967 manager.go:334] "Starting Device Plugin manager" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.199806 4967 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.199838 4967 server.go:79] "Starting device plugin registration server" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.200615 4967 eviction_manager.go:189] "Eviction manager: starting control loop" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.200659 4967 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.200999 4967 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.201083 4967 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.201091 4967 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Oct 08 00:06:42 crc kubenswrapper[4967]: E1008 00:06:42.213876 4967 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.217178 4967 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.217374 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.218742 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.218779 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.218793 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.218958 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.219187 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.219255 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.219880 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.219929 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.219940 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.220103 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.220264 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.220313 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.220476 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.220509 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.220522 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.221089 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.221120 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.221132 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.221202 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.221219 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.221229 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.221340 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.221495 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.221532 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.222020 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.222057 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.222069 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.222187 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.222326 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.222362 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.222441 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.222469 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.222482 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.223210 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.223231 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.223239 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.223276 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.223294 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.223305 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.223385 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.223405 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.224482 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.224508 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.224515 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:06:42 crc kubenswrapper[4967]: E1008 00:06:42.246981 4967 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.20:6443: connect: connection refused" interval="400ms" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.275338 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.275383 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.275407 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.275428 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.275449 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.275473 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.275496 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.275514 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.275532 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.275552 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.275571 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.275590 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.275610 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.275641 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.275661 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.301021 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.302445 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.302486 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.302497 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.302523 4967 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 08 00:06:42 crc kubenswrapper[4967]: E1008 00:06:42.303840 4967 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.20:6443: connect: connection refused" node="crc" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.377082 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.377157 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.377232 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.377266 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.377295 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.377325 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.377353 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.377382 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.377380 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.377414 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.377434 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.377443 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.377402 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.377500 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.377485 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.377545 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.377587 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.377534 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.377598 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.377548 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.377625 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.377376 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.377632 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.377746 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.377755 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.377832 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.377833 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.377881 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.377945 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.378059 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.504303 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.506089 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.506149 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.506168 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.506219 4967 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 08 00:06:42 crc kubenswrapper[4967]: E1008 00:06:42.506875 4967 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.20:6443: connect: connection refused" node="crc" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.557382 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.566444 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 00:06:42 crc kubenswrapper[4967]: E1008 00:06:42.566606 4967 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.20:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.186c5b5b4a3b1af8 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-10-08 00:06:42.040560376 +0000 UTC m=+0.691095035,LastTimestamp:2025-10-08 00:06:42.040560376 +0000 UTC m=+0.691095035,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.594053 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.620623 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.628991 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 08 00:06:42 crc kubenswrapper[4967]: W1008 00:06:42.644710 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-99203311124cb92ec3304c76f20d8895d8a35b6cefc8f2a46e60d543f413d330 WatchSource:0}: Error finding container 99203311124cb92ec3304c76f20d8895d8a35b6cefc8f2a46e60d543f413d330: Status 404 returned error can't find the container with id 99203311124cb92ec3304c76f20d8895d8a35b6cefc8f2a46e60d543f413d330 Oct 08 00:06:42 crc kubenswrapper[4967]: E1008 00:06:42.647865 4967 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.20:6443: connect: connection refused" interval="800ms" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.907838 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.909640 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.909708 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.909721 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:06:42 crc kubenswrapper[4967]: I1008 00:06:42.909747 4967 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 08 00:06:42 crc kubenswrapper[4967]: E1008 00:06:42.910236 4967 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.20:6443: connect: connection refused" node="crc" Oct 08 00:06:42 crc kubenswrapper[4967]: W1008 00:06:42.995020 4967 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.20:6443: connect: connection refused Oct 08 00:06:42 crc kubenswrapper[4967]: E1008 00:06:42.995146 4967 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.20:6443: connect: connection refused" logger="UnhandledError" Oct 08 00:06:43 crc kubenswrapper[4967]: W1008 00:06:43.028402 4967 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.20:6443: connect: connection refused Oct 08 00:06:43 crc kubenswrapper[4967]: E1008 00:06:43.028521 4967 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.20:6443: connect: connection refused" logger="UnhandledError" Oct 08 00:06:43 crc kubenswrapper[4967]: I1008 00:06:43.043463 4967 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.20:6443: connect: connection refused Oct 08 00:06:43 crc kubenswrapper[4967]: W1008 00:06:43.051512 4967 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.20:6443: connect: connection refused Oct 08 00:06:43 crc kubenswrapper[4967]: E1008 00:06:43.051588 4967 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.20:6443: connect: connection refused" logger="UnhandledError" Oct 08 00:06:43 crc kubenswrapper[4967]: I1008 00:06:43.122106 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"58b0fc5d28a284727a5d0ec0a4ee6b4e49255461ca8269047941d169c9be27be"} Oct 08 00:06:43 crc kubenswrapper[4967]: I1008 00:06:43.123509 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"63b2318aa9bdbd5d80219b2c57c4e4b06754242da6ae2a6f3d1f3507ce31c705"} Oct 08 00:06:43 crc kubenswrapper[4967]: I1008 00:06:43.124763 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"99203311124cb92ec3304c76f20d8895d8a35b6cefc8f2a46e60d543f413d330"} Oct 08 00:06:43 crc kubenswrapper[4967]: I1008 00:06:43.126242 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"1a8908f4fb55a84307f27932974a9a9e161871b3e74ffc3541d36d3e57e3206d"} Oct 08 00:06:43 crc kubenswrapper[4967]: I1008 00:06:43.127150 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"a0c0c74bb1e258ba42e867f8a09366c2d864f4cb66bbd68b44b72817dc25d57d"} Oct 08 00:06:43 crc kubenswrapper[4967]: E1008 00:06:43.449047 4967 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.20:6443: connect: connection refused" interval="1.6s" Oct 08 00:06:43 crc kubenswrapper[4967]: W1008 00:06:43.538821 4967 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.20:6443: connect: connection refused Oct 08 00:06:43 crc kubenswrapper[4967]: E1008 00:06:43.538953 4967 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.20:6443: connect: connection refused" logger="UnhandledError" Oct 08 00:06:43 crc kubenswrapper[4967]: I1008 00:06:43.711185 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 00:06:43 crc kubenswrapper[4967]: I1008 00:06:43.713048 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:06:43 crc kubenswrapper[4967]: I1008 00:06:43.713090 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:06:43 crc kubenswrapper[4967]: I1008 00:06:43.713106 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:06:43 crc kubenswrapper[4967]: I1008 00:06:43.713140 4967 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 08 00:06:43 crc kubenswrapper[4967]: E1008 00:06:43.713709 4967 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.20:6443: connect: connection refused" node="crc" Oct 08 00:06:44 crc kubenswrapper[4967]: I1008 00:06:44.044069 4967 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.20:6443: connect: connection refused Oct 08 00:06:44 crc kubenswrapper[4967]: I1008 00:06:44.139819 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"6aa6f5e81feceab3d1c7587152a70546b365e629a69267e4d3084324a1de764a"} Oct 08 00:06:44 crc kubenswrapper[4967]: I1008 00:06:44.139891 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"a22f5dabb39da0a7bd01db4dae1437b9a231a1fcc0dfe749a4a5a637a72cf33c"} Oct 08 00:06:44 crc kubenswrapper[4967]: I1008 00:06:44.139906 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"01a2b3dc785520c79eae998482e5a1347725c76584357073238fe60a619edb77"} Oct 08 00:06:44 crc kubenswrapper[4967]: I1008 00:06:44.142682 4967 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="b5b222ad84c31d8746dfe03124c7287b401deb3ca2c62ce2b4fbd40e638cee0e" exitCode=0 Oct 08 00:06:44 crc kubenswrapper[4967]: I1008 00:06:44.142774 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"b5b222ad84c31d8746dfe03124c7287b401deb3ca2c62ce2b4fbd40e638cee0e"} Oct 08 00:06:44 crc kubenswrapper[4967]: I1008 00:06:44.142807 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 00:06:44 crc kubenswrapper[4967]: I1008 00:06:44.143651 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:06:44 crc kubenswrapper[4967]: I1008 00:06:44.143693 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:06:44 crc kubenswrapper[4967]: I1008 00:06:44.143706 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:06:44 crc kubenswrapper[4967]: I1008 00:06:44.146222 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 00:06:44 crc kubenswrapper[4967]: I1008 00:06:44.146753 4967 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="e7fd2ff88cd8fe658d0c7c78edd6ac27b3d39cbcb6984dc7430c9696acc57c72" exitCode=0 Oct 08 00:06:44 crc kubenswrapper[4967]: I1008 00:06:44.146795 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"e7fd2ff88cd8fe658d0c7c78edd6ac27b3d39cbcb6984dc7430c9696acc57c72"} Oct 08 00:06:44 crc kubenswrapper[4967]: I1008 00:06:44.146955 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 00:06:44 crc kubenswrapper[4967]: I1008 00:06:44.147740 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:06:44 crc kubenswrapper[4967]: I1008 00:06:44.147772 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:06:44 crc kubenswrapper[4967]: I1008 00:06:44.147786 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:06:44 crc kubenswrapper[4967]: I1008 00:06:44.148580 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:06:44 crc kubenswrapper[4967]: I1008 00:06:44.148786 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:06:44 crc kubenswrapper[4967]: I1008 00:06:44.149041 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:06:44 crc kubenswrapper[4967]: I1008 00:06:44.150838 4967 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="ae14677731c53e87871129d2140a34b18f0a48555309b652ebd490996fb1a723" exitCode=0 Oct 08 00:06:44 crc kubenswrapper[4967]: I1008 00:06:44.150882 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"ae14677731c53e87871129d2140a34b18f0a48555309b652ebd490996fb1a723"} Oct 08 00:06:44 crc kubenswrapper[4967]: I1008 00:06:44.150955 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 00:06:44 crc kubenswrapper[4967]: I1008 00:06:44.152140 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:06:44 crc kubenswrapper[4967]: I1008 00:06:44.152190 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:06:44 crc kubenswrapper[4967]: I1008 00:06:44.152204 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:06:44 crc kubenswrapper[4967]: I1008 00:06:44.153609 4967 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="426e86ab409776ff0fa9505dc0d8ec36f27a5bf21d3aedafb19d3935bdba7529" exitCode=0 Oct 08 00:06:44 crc kubenswrapper[4967]: I1008 00:06:44.153661 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"426e86ab409776ff0fa9505dc0d8ec36f27a5bf21d3aedafb19d3935bdba7529"} Oct 08 00:06:44 crc kubenswrapper[4967]: I1008 00:06:44.154128 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 00:06:44 crc kubenswrapper[4967]: I1008 00:06:44.156009 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:06:44 crc kubenswrapper[4967]: I1008 00:06:44.156174 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:06:44 crc kubenswrapper[4967]: I1008 00:06:44.156311 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:06:45 crc kubenswrapper[4967]: I1008 00:06:45.043443 4967 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.20:6443: connect: connection refused Oct 08 00:06:45 crc kubenswrapper[4967]: E1008 00:06:45.050357 4967 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.20:6443: connect: connection refused" interval="3.2s" Oct 08 00:06:45 crc kubenswrapper[4967]: I1008 00:06:45.160793 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"7ca28d8db19036e47a7ef3dadb70941f6c6cf5735e7703977d036185c46fce40"} Oct 08 00:06:45 crc kubenswrapper[4967]: I1008 00:06:45.160859 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"9d0a0b2ef4a98ee828688081a74f4b7ecc31c8fae415d7257a560ecd43960ce3"} Oct 08 00:06:45 crc kubenswrapper[4967]: I1008 00:06:45.160884 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"b96b53b1b283025b3d078a3f505d469d1872ad4bb8ab8e612c36c70b0b00a920"} Oct 08 00:06:45 crc kubenswrapper[4967]: I1008 00:06:45.160903 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"fbadcdade274b6d9be940f22bc949e3c55cadf95dcad43c8cb7a1085c0e6708c"} Oct 08 00:06:45 crc kubenswrapper[4967]: I1008 00:06:45.164062 4967 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="84e8f87487a6681581394dec7d205954810f0a473457a349f324845fdbdc16d1" exitCode=0 Oct 08 00:06:45 crc kubenswrapper[4967]: I1008 00:06:45.164155 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"84e8f87487a6681581394dec7d205954810f0a473457a349f324845fdbdc16d1"} Oct 08 00:06:45 crc kubenswrapper[4967]: I1008 00:06:45.164201 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 00:06:45 crc kubenswrapper[4967]: I1008 00:06:45.165148 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:06:45 crc kubenswrapper[4967]: I1008 00:06:45.165177 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:06:45 crc kubenswrapper[4967]: I1008 00:06:45.165188 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:06:45 crc kubenswrapper[4967]: I1008 00:06:45.168637 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"acfae51bf8bee1b379a3859b7fb98ad0df3bed5d49dc0abe82b917512955c009"} Oct 08 00:06:45 crc kubenswrapper[4967]: I1008 00:06:45.168790 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 00:06:45 crc kubenswrapper[4967]: I1008 00:06:45.171762 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:06:45 crc kubenswrapper[4967]: I1008 00:06:45.171809 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:06:45 crc kubenswrapper[4967]: I1008 00:06:45.171825 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:06:45 crc kubenswrapper[4967]: I1008 00:06:45.174971 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"c430979d298477b4394c3567bb3800ab7dea874cee8fa41a6a50a19a7e3a0133"} Oct 08 00:06:45 crc kubenswrapper[4967]: I1008 00:06:45.175019 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"9c7893dfaa0370f159d41c30cd869bb18550a819e8c3d20dd6f57dcbcc697a57"} Oct 08 00:06:45 crc kubenswrapper[4967]: I1008 00:06:45.175033 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"37fd5dcdfb864320d4d17729ab48c6da505fd867f384ad70a6dcf5bed78a3978"} Oct 08 00:06:45 crc kubenswrapper[4967]: I1008 00:06:45.174991 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 00:06:45 crc kubenswrapper[4967]: I1008 00:06:45.177626 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"63c2b7c4b2791bd1f16888c58b878dbff1f87884e68c9949de914d0d3e1d39fb"} Oct 08 00:06:45 crc kubenswrapper[4967]: I1008 00:06:45.177788 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 00:06:45 crc kubenswrapper[4967]: I1008 00:06:45.179907 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:06:45 crc kubenswrapper[4967]: I1008 00:06:45.179943 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:06:45 crc kubenswrapper[4967]: I1008 00:06:45.179954 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:06:45 crc kubenswrapper[4967]: I1008 00:06:45.181356 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:06:45 crc kubenswrapper[4967]: I1008 00:06:45.181406 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:06:45 crc kubenswrapper[4967]: I1008 00:06:45.181421 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:06:45 crc kubenswrapper[4967]: I1008 00:06:45.314411 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 00:06:45 crc kubenswrapper[4967]: I1008 00:06:45.316174 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:06:45 crc kubenswrapper[4967]: I1008 00:06:45.316229 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:06:45 crc kubenswrapper[4967]: I1008 00:06:45.316240 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:06:45 crc kubenswrapper[4967]: I1008 00:06:45.316271 4967 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 08 00:06:45 crc kubenswrapper[4967]: E1008 00:06:45.316801 4967 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.20:6443: connect: connection refused" node="crc" Oct 08 00:06:45 crc kubenswrapper[4967]: W1008 00:06:45.552773 4967 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.20:6443: connect: connection refused Oct 08 00:06:45 crc kubenswrapper[4967]: E1008 00:06:45.552911 4967 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.20:6443: connect: connection refused" logger="UnhandledError" Oct 08 00:06:45 crc kubenswrapper[4967]: W1008 00:06:45.831013 4967 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.20:6443: connect: connection refused Oct 08 00:06:45 crc kubenswrapper[4967]: E1008 00:06:45.831118 4967 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.20:6443: connect: connection refused" logger="UnhandledError" Oct 08 00:06:46 crc kubenswrapper[4967]: I1008 00:06:46.043862 4967 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.20:6443: connect: connection refused Oct 08 00:06:46 crc kubenswrapper[4967]: W1008 00:06:46.119398 4967 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.20:6443: connect: connection refused Oct 08 00:06:46 crc kubenswrapper[4967]: E1008 00:06:46.119541 4967 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.20:6443: connect: connection refused" logger="UnhandledError" Oct 08 00:06:46 crc kubenswrapper[4967]: I1008 00:06:46.182635 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 08 00:06:46 crc kubenswrapper[4967]: I1008 00:06:46.185302 4967 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="2c3118c3b932a91a617645840254d372bca2799d4b08eb2b371f5d2a0a6e116d" exitCode=255 Oct 08 00:06:46 crc kubenswrapper[4967]: I1008 00:06:46.185398 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"2c3118c3b932a91a617645840254d372bca2799d4b08eb2b371f5d2a0a6e116d"} Oct 08 00:06:46 crc kubenswrapper[4967]: I1008 00:06:46.185475 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 00:06:46 crc kubenswrapper[4967]: I1008 00:06:46.187384 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:06:46 crc kubenswrapper[4967]: I1008 00:06:46.187432 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:06:46 crc kubenswrapper[4967]: I1008 00:06:46.187449 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:06:46 crc kubenswrapper[4967]: I1008 00:06:46.188285 4967 scope.go:117] "RemoveContainer" containerID="2c3118c3b932a91a617645840254d372bca2799d4b08eb2b371f5d2a0a6e116d" Oct 08 00:06:46 crc kubenswrapper[4967]: I1008 00:06:46.189057 4967 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="94ca32f1ff66c8c2967bcc4ae1923e8d1792a9bdfb14e5fc9562b79ebd4c4b31" exitCode=0 Oct 08 00:06:46 crc kubenswrapper[4967]: I1008 00:06:46.189162 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 00:06:46 crc kubenswrapper[4967]: I1008 00:06:46.189183 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 00:06:46 crc kubenswrapper[4967]: I1008 00:06:46.189232 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"94ca32f1ff66c8c2967bcc4ae1923e8d1792a9bdfb14e5fc9562b79ebd4c4b31"} Oct 08 00:06:46 crc kubenswrapper[4967]: I1008 00:06:46.189283 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 00:06:46 crc kubenswrapper[4967]: I1008 00:06:46.189365 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 00:06:46 crc kubenswrapper[4967]: I1008 00:06:46.189384 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 08 00:06:46 crc kubenswrapper[4967]: I1008 00:06:46.190630 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:06:46 crc kubenswrapper[4967]: I1008 00:06:46.190662 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:06:46 crc kubenswrapper[4967]: I1008 00:06:46.190692 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:06:46 crc kubenswrapper[4967]: I1008 00:06:46.190727 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:06:46 crc kubenswrapper[4967]: I1008 00:06:46.190771 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:06:46 crc kubenswrapper[4967]: I1008 00:06:46.190793 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:06:46 crc kubenswrapper[4967]: I1008 00:06:46.191047 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:06:46 crc kubenswrapper[4967]: I1008 00:06:46.191089 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:06:46 crc kubenswrapper[4967]: I1008 00:06:46.191111 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:06:46 crc kubenswrapper[4967]: I1008 00:06:46.191480 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:06:46 crc kubenswrapper[4967]: I1008 00:06:46.191518 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:06:46 crc kubenswrapper[4967]: I1008 00:06:46.191540 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:06:46 crc kubenswrapper[4967]: I1008 00:06:46.542386 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 00:06:46 crc kubenswrapper[4967]: I1008 00:06:46.932862 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 00:06:47 crc kubenswrapper[4967]: I1008 00:06:47.194175 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 08 00:06:47 crc kubenswrapper[4967]: I1008 00:06:47.196375 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"1400e110997c64fa33d6b8672c7e91cab5f782952ddf8e1dea2a882689026e7f"} Oct 08 00:06:47 crc kubenswrapper[4967]: I1008 00:06:47.196494 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 00:06:47 crc kubenswrapper[4967]: I1008 00:06:47.197947 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:06:47 crc kubenswrapper[4967]: I1008 00:06:47.197986 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:06:47 crc kubenswrapper[4967]: I1008 00:06:47.197999 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:06:47 crc kubenswrapper[4967]: I1008 00:06:47.202099 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"3caa55b6d3dcc51f7e25678ac48788a15cb24e3e3bfa2ed03537e1aa9b50036f"} Oct 08 00:06:47 crc kubenswrapper[4967]: I1008 00:06:47.202158 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"4d6a67f4b5d47cd7d68ac434c686ac3b35028c0cb926649e5319e1094c0186dc"} Oct 08 00:06:47 crc kubenswrapper[4967]: I1008 00:06:47.202184 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"4043e9f8b86926db8f4224c526ac599ece4334db22f92ccecad393f39713d213"} Oct 08 00:06:47 crc kubenswrapper[4967]: I1008 00:06:47.202208 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"34f33ad9f968c14b24b17b5f4dd2fae884afe360fc501cbc704b31d5ee5d0e38"} Oct 08 00:06:47 crc kubenswrapper[4967]: I1008 00:06:47.202226 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 00:06:47 crc kubenswrapper[4967]: I1008 00:06:47.203249 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:06:47 crc kubenswrapper[4967]: I1008 00:06:47.203298 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:06:47 crc kubenswrapper[4967]: I1008 00:06:47.203315 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:06:47 crc kubenswrapper[4967]: I1008 00:06:47.860279 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 00:06:47 crc kubenswrapper[4967]: I1008 00:06:47.860620 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 00:06:47 crc kubenswrapper[4967]: I1008 00:06:47.862614 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:06:47 crc kubenswrapper[4967]: I1008 00:06:47.862705 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:06:47 crc kubenswrapper[4967]: I1008 00:06:47.862723 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:06:47 crc kubenswrapper[4967]: I1008 00:06:47.869466 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 00:06:48 crc kubenswrapper[4967]: I1008 00:06:48.211406 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"18725109451a19f37f978daaa770a547cb3f1ae32bf3ee7f9f9789df0d2ba43f"} Oct 08 00:06:48 crc kubenswrapper[4967]: I1008 00:06:48.211450 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 00:06:48 crc kubenswrapper[4967]: I1008 00:06:48.211581 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 00:06:48 crc kubenswrapper[4967]: I1008 00:06:48.211618 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 00:06:48 crc kubenswrapper[4967]: I1008 00:06:48.211762 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 00:06:48 crc kubenswrapper[4967]: I1008 00:06:48.213127 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:06:48 crc kubenswrapper[4967]: I1008 00:06:48.213169 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:06:48 crc kubenswrapper[4967]: I1008 00:06:48.213182 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:06:48 crc kubenswrapper[4967]: I1008 00:06:48.213248 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:06:48 crc kubenswrapper[4967]: I1008 00:06:48.213293 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:06:48 crc kubenswrapper[4967]: I1008 00:06:48.213315 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:06:48 crc kubenswrapper[4967]: I1008 00:06:48.213252 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:06:48 crc kubenswrapper[4967]: I1008 00:06:48.213515 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:06:48 crc kubenswrapper[4967]: I1008 00:06:48.213542 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:06:48 crc kubenswrapper[4967]: I1008 00:06:48.517787 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 00:06:48 crc kubenswrapper[4967]: I1008 00:06:48.519722 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:06:48 crc kubenswrapper[4967]: I1008 00:06:48.519787 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:06:48 crc kubenswrapper[4967]: I1008 00:06:48.519806 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:06:48 crc kubenswrapper[4967]: I1008 00:06:48.519843 4967 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 08 00:06:49 crc kubenswrapper[4967]: I1008 00:06:49.214039 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 00:06:49 crc kubenswrapper[4967]: I1008 00:06:49.214039 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 00:06:49 crc kubenswrapper[4967]: I1008 00:06:49.215141 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 00:06:49 crc kubenswrapper[4967]: I1008 00:06:49.215283 4967 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 08 00:06:49 crc kubenswrapper[4967]: I1008 00:06:49.215382 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 00:06:49 crc kubenswrapper[4967]: I1008 00:06:49.215634 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:06:49 crc kubenswrapper[4967]: I1008 00:06:49.215711 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:06:49 crc kubenswrapper[4967]: I1008 00:06:49.215730 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:06:49 crc kubenswrapper[4967]: I1008 00:06:49.215859 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:06:49 crc kubenswrapper[4967]: I1008 00:06:49.215903 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:06:49 crc kubenswrapper[4967]: I1008 00:06:49.215920 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:06:49 crc kubenswrapper[4967]: I1008 00:06:49.216619 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:06:49 crc kubenswrapper[4967]: I1008 00:06:49.216698 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:06:49 crc kubenswrapper[4967]: I1008 00:06:49.216723 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:06:49 crc kubenswrapper[4967]: I1008 00:06:49.747337 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 00:06:50 crc kubenswrapper[4967]: I1008 00:06:50.219428 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 00:06:50 crc kubenswrapper[4967]: I1008 00:06:50.220821 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:06:50 crc kubenswrapper[4967]: I1008 00:06:50.220887 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:06:50 crc kubenswrapper[4967]: I1008 00:06:50.220917 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:06:50 crc kubenswrapper[4967]: I1008 00:06:50.458776 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 00:06:50 crc kubenswrapper[4967]: I1008 00:06:50.458981 4967 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 08 00:06:50 crc kubenswrapper[4967]: I1008 00:06:50.459042 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 00:06:50 crc kubenswrapper[4967]: I1008 00:06:50.460729 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:06:50 crc kubenswrapper[4967]: I1008 00:06:50.460796 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:06:50 crc kubenswrapper[4967]: I1008 00:06:50.460816 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:06:51 crc kubenswrapper[4967]: I1008 00:06:51.650204 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Oct 08 00:06:51 crc kubenswrapper[4967]: I1008 00:06:51.650536 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 00:06:51 crc kubenswrapper[4967]: I1008 00:06:51.655880 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:06:51 crc kubenswrapper[4967]: I1008 00:06:51.655948 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:06:51 crc kubenswrapper[4967]: I1008 00:06:51.655968 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:06:51 crc kubenswrapper[4967]: I1008 00:06:51.930788 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 00:06:51 crc kubenswrapper[4967]: I1008 00:06:51.931054 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 00:06:51 crc kubenswrapper[4967]: I1008 00:06:51.932768 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:06:51 crc kubenswrapper[4967]: I1008 00:06:51.932821 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:06:51 crc kubenswrapper[4967]: I1008 00:06:51.932842 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:06:52 crc kubenswrapper[4967]: E1008 00:06:52.214431 4967 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 08 00:06:52 crc kubenswrapper[4967]: I1008 00:06:52.276294 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Oct 08 00:06:52 crc kubenswrapper[4967]: I1008 00:06:52.276516 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 00:06:52 crc kubenswrapper[4967]: I1008 00:06:52.278047 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:06:52 crc kubenswrapper[4967]: I1008 00:06:52.278097 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:06:52 crc kubenswrapper[4967]: I1008 00:06:52.278117 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:06:53 crc kubenswrapper[4967]: I1008 00:06:53.459750 4967 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 08 00:06:53 crc kubenswrapper[4967]: I1008 00:06:53.459891 4967 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 08 00:06:56 crc kubenswrapper[4967]: W1008 00:06:56.222387 4967 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout Oct 08 00:06:56 crc kubenswrapper[4967]: I1008 00:06:56.222530 4967 trace.go:236] Trace[1383959825]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (08-Oct-2025 00:06:46.220) (total time: 10002ms): Oct 08 00:06:56 crc kubenswrapper[4967]: Trace[1383959825]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10002ms (00:06:56.222) Oct 08 00:06:56 crc kubenswrapper[4967]: Trace[1383959825]: [10.002149012s] [10.002149012s] END Oct 08 00:06:56 crc kubenswrapper[4967]: E1008 00:06:56.222572 4967 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Oct 08 00:06:57 crc kubenswrapper[4967]: I1008 00:06:57.044636 4967 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Oct 08 00:06:57 crc kubenswrapper[4967]: I1008 00:06:57.291390 4967 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 08 00:06:57 crc kubenswrapper[4967]: I1008 00:06:57.291482 4967 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 08 00:06:57 crc kubenswrapper[4967]: I1008 00:06:57.299143 4967 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 08 00:06:57 crc kubenswrapper[4967]: I1008 00:06:57.299239 4967 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 08 00:06:59 crc kubenswrapper[4967]: I1008 00:06:59.755646 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 00:06:59 crc kubenswrapper[4967]: I1008 00:06:59.755924 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 00:06:59 crc kubenswrapper[4967]: I1008 00:06:59.757452 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:06:59 crc kubenswrapper[4967]: I1008 00:06:59.757509 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:06:59 crc kubenswrapper[4967]: I1008 00:06:59.757528 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:06:59 crc kubenswrapper[4967]: I1008 00:06:59.763516 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 00:07:00 crc kubenswrapper[4967]: I1008 00:07:00.248075 4967 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 08 00:07:00 crc kubenswrapper[4967]: I1008 00:07:00.248149 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 00:07:00 crc kubenswrapper[4967]: I1008 00:07:00.249446 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:00 crc kubenswrapper[4967]: I1008 00:07:00.249526 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:00 crc kubenswrapper[4967]: I1008 00:07:00.249576 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:01 crc kubenswrapper[4967]: I1008 00:07:01.697627 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Oct 08 00:07:01 crc kubenswrapper[4967]: I1008 00:07:01.697850 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 00:07:01 crc kubenswrapper[4967]: I1008 00:07:01.699579 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:01 crc kubenswrapper[4967]: I1008 00:07:01.699650 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:01 crc kubenswrapper[4967]: I1008 00:07:01.699720 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:01 crc kubenswrapper[4967]: I1008 00:07:01.717088 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Oct 08 00:07:01 crc kubenswrapper[4967]: I1008 00:07:01.937807 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 00:07:01 crc kubenswrapper[4967]: I1008 00:07:01.938011 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 00:07:01 crc kubenswrapper[4967]: I1008 00:07:01.939232 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:01 crc kubenswrapper[4967]: I1008 00:07:01.939253 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:01 crc kubenswrapper[4967]: I1008 00:07:01.939282 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:02 crc kubenswrapper[4967]: E1008 00:07:02.215483 4967 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 08 00:07:02 crc kubenswrapper[4967]: I1008 00:07:02.253387 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 00:07:02 crc kubenswrapper[4967]: I1008 00:07:02.254619 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:02 crc kubenswrapper[4967]: I1008 00:07:02.254660 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:02 crc kubenswrapper[4967]: I1008 00:07:02.254690 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:02 crc kubenswrapper[4967]: E1008 00:07:02.287138 4967 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Oct 08 00:07:02 crc kubenswrapper[4967]: I1008 00:07:02.292382 4967 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Oct 08 00:07:02 crc kubenswrapper[4967]: I1008 00:07:02.292466 4967 trace.go:236] Trace[1831890304]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (08-Oct-2025 00:06:51.618) (total time: 10673ms): Oct 08 00:07:02 crc kubenswrapper[4967]: Trace[1831890304]: ---"Objects listed" error: 10673ms (00:07:02.292) Oct 08 00:07:02 crc kubenswrapper[4967]: Trace[1831890304]: [10.673603252s] [10.673603252s] END Oct 08 00:07:02 crc kubenswrapper[4967]: I1008 00:07:02.292500 4967 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Oct 08 00:07:02 crc kubenswrapper[4967]: I1008 00:07:02.295043 4967 trace.go:236] Trace[702480006]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (08-Oct-2025 00:06:50.893) (total time: 11401ms): Oct 08 00:07:02 crc kubenswrapper[4967]: Trace[702480006]: ---"Objects listed" error: 11401ms (00:07:02.294) Oct 08 00:07:02 crc kubenswrapper[4967]: Trace[702480006]: [11.40193539s] [11.40193539s] END Oct 08 00:07:02 crc kubenswrapper[4967]: I1008 00:07:02.295080 4967 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Oct 08 00:07:02 crc kubenswrapper[4967]: I1008 00:07:02.297080 4967 trace.go:236] Trace[46873440]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (08-Oct-2025 00:06:48.830) (total time: 13466ms): Oct 08 00:07:02 crc kubenswrapper[4967]: Trace[46873440]: ---"Objects listed" error: 13466ms (00:07:02.296) Oct 08 00:07:02 crc kubenswrapper[4967]: Trace[46873440]: [13.466140462s] [13.466140462s] END Oct 08 00:07:02 crc kubenswrapper[4967]: I1008 00:07:02.297104 4967 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Oct 08 00:07:02 crc kubenswrapper[4967]: E1008 00:07:02.298006 4967 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Oct 08 00:07:02 crc kubenswrapper[4967]: I1008 00:07:02.302143 4967 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Oct 08 00:07:02 crc kubenswrapper[4967]: I1008 00:07:02.351165 4967 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:37178->192.168.126.11:17697: read: connection reset by peer" start-of-body= Oct 08 00:07:02 crc kubenswrapper[4967]: I1008 00:07:02.351250 4967 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:37178->192.168.126.11:17697: read: connection reset by peer" Oct 08 00:07:02 crc kubenswrapper[4967]: I1008 00:07:02.351708 4967 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Oct 08 00:07:02 crc kubenswrapper[4967]: I1008 00:07:02.351770 4967 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Oct 08 00:07:02 crc kubenswrapper[4967]: I1008 00:07:02.781746 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 00:07:02 crc kubenswrapper[4967]: I1008 00:07:02.785299 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.037829 4967 apiserver.go:52] "Watching apiserver" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.043490 4967 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.044265 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"] Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.045483 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.045770 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.045921 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 00:07:03 crc kubenswrapper[4967]: E1008 00:07:03.046013 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 00:07:03 crc kubenswrapper[4967]: E1008 00:07:03.046157 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.046394 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.046620 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.046747 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 00:07:03 crc kubenswrapper[4967]: E1008 00:07:03.046840 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.051573 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.051583 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.052340 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.052370 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.052356 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.052704 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.052815 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.053047 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.053069 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.096524 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.096588 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.096626 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.097439 4967 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.103698 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.123992 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.129232 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.146114 4967 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.152008 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.166870 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.177923 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.188578 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.197326 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.197420 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.197476 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.197525 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.197580 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.197631 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.197710 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.197761 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.197809 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.197861 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.197907 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.197957 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.197986 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.198004 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.198113 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.198176 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.198226 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.198273 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.198317 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.198367 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.198411 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.198453 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.198494 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.198538 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.198581 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.198627 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.198713 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.198766 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.198811 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.198864 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.198912 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.198956 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.199002 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.199131 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.199219 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.199264 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.199316 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.199371 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.199419 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.199475 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.199517 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.199555 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.199595 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.199639 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.199750 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.199799 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.199842 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.199891 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.199933 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.199973 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.200018 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.200064 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.200109 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.200161 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.200210 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.200312 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.200362 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.200451 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.200504 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.200554 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.200603 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.200649 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.200736 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.200788 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.200840 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.200889 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.200939 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.200984 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.201031 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.201086 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.201140 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.201185 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.201235 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.201281 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.201327 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.201373 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.198091 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.198404 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.201423 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.201473 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.201517 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.201563 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.201622 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.201858 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.201913 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.201959 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.202012 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.202067 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.202113 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.202157 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.202200 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.202248 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.202295 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.202400 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.202576 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.202635 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.202724 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.202773 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.202823 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.202874 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.202923 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.202968 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.203016 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.203065 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.203111 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.203155 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.203199 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.203255 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.203310 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.203361 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.203410 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.203464 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.203517 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.203568 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.203624 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.203747 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.203807 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.203857 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.203914 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.203972 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.204027 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.204084 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.204153 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.204206 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.204257 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.204304 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.204351 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.204407 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.204467 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.204524 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.204579 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.204639 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.204728 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.204783 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.204835 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.204891 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.204941 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.204999 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.205057 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.205112 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.205164 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.205217 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.205280 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.205335 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.205389 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.205440 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.205498 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.205551 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.205606 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.205660 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.206904 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.206968 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.207021 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.207073 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.207124 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.207181 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.207233 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.207280 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.207329 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.207376 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.207424 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.207477 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.207532 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.207590 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.207640 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.207764 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.207822 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.207883 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.207938 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.207990 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.208041 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.208088 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.208138 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.208192 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.208244 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.208307 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.208358 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.208409 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.208461 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.208513 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.208566 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.208619 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.208720 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.208789 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.208844 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.208898 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.208954 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.209009 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.209061 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.209116 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.209170 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.209225 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.209277 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.209331 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.209386 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.209443 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.209500 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.209554 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.209699 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.209784 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.209845 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.209948 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.210009 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.210069 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.210128 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.210187 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.210243 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.210297 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.210351 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.210441 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.210578 4967 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.210612 4967 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.210642 4967 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.198804 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.218301 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.199591 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.200054 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.200072 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.200258 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.200462 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.200803 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.201383 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.201513 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.201625 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.201929 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.202049 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.202817 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.203083 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.203206 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.203339 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.203369 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.203307 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.203745 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.203795 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.203801 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.203971 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.204180 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.204178 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.204377 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.204380 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.204943 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.204951 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.205277 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.205505 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.205790 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.206392 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.206543 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.218911 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.207516 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.208595 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.208707 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.209087 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.209455 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: E1008 00:07:03.210851 4967 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.212966 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.213175 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.214099 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.214715 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.214764 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.215774 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.217305 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.217353 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.217621 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.217902 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.219480 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.219612 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.219758 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.219866 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.220235 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.220299 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.220338 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.220872 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.221338 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.220711 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.221577 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.221756 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.221793 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.222088 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.222258 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.222618 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.222788 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.222927 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.222962 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.223151 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.223343 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.223371 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.223416 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.223902 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.224201 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.224301 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.224321 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.224512 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.224612 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.224817 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.224980 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.225449 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.225725 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.225894 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.226064 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.226370 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.226432 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.226765 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.227709 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.228085 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.230049 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.230302 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.231234 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.231647 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.231714 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.232027 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.232300 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.232741 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.233145 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.236455 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.237437 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.237945 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.237980 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.237983 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.238346 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.238647 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.239180 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: E1008 00:07:03.239324 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 00:07:03.739293966 +0000 UTC m=+22.389828615 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:07:03 crc kubenswrapper[4967]: E1008 00:07:03.239437 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 00:07:03.73942889 +0000 UTC m=+22.389963539 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.240379 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.240605 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.240895 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.241412 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.218255 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.241844 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.242038 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.243127 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.243366 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.243510 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.244258 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.244390 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.244656 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.244810 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.245306 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.245338 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.245687 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.245897 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.246160 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.246656 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.249209 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.249276 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.249632 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.249906 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.250019 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.250253 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.250695 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.250815 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.250883 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.251001 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.251209 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.251305 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.252403 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.252533 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.252643 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.253256 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.253881 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.254264 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: E1008 00:07:03.254349 4967 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 00:07:03 crc kubenswrapper[4967]: E1008 00:07:03.254431 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 00:07:03.754418829 +0000 UTC m=+22.404953478 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.254513 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.254944 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.255162 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.255202 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.255295 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.255841 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.256612 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.257307 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.258559 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.258937 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.259075 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.260593 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.261220 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.261864 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.262359 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.262160 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.262396 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.278420 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.278590 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.281024 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.281353 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.281400 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.281693 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.282152 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.282223 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.282302 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.282340 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.282847 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.283222 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.283231 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.285180 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58c4704-a88a-43b4-b944-4de626f13f7e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22f5dabb39da0a7bd01db4dae1437b9a231a1fcc0dfe749a4a5a637a72cf33c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01a2b3dc785520c79eae998482e5a1347725c76584357073238fe60a619edb77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6aa6f5e81feceab3d1c7587152a70546b365e629a69267e4d3084324a1de764a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63c2b7c4b2791bd1f16888c58b878dbff1f87884e68c9949de914d0d3e1d39fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.285486 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.285592 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: E1008 00:07:03.285556 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 00:07:03 crc kubenswrapper[4967]: E1008 00:07:03.285712 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 00:07:03 crc kubenswrapper[4967]: E1008 00:07:03.285747 4967 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 00:07:03 crc kubenswrapper[4967]: E1008 00:07:03.285870 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-08 00:07:03.785833615 +0000 UTC m=+22.436368284 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.285917 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.285986 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.286136 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.286381 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.286721 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.287377 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.287758 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.287816 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.288734 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.289529 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.290583 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.290707 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.290742 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.290963 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.293167 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.300933 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.302108 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 08 00:07:03 crc kubenswrapper[4967]: E1008 00:07:03.311006 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 00:07:03 crc kubenswrapper[4967]: E1008 00:07:03.311043 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 00:07:03 crc kubenswrapper[4967]: E1008 00:07:03.311061 4967 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 00:07:03 crc kubenswrapper[4967]: E1008 00:07:03.311131 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-08 00:07:03.811106683 +0000 UTC m=+22.461641332 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.311280 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.311461 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.311834 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.311944 4967 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.311947 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.311965 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.312009 4967 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.312028 4967 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.312042 4967 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.312057 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.312071 4967 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.312084 4967 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.312096 4967 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.312109 4967 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.312121 4967 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.312132 4967 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.312171 4967 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.312185 4967 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.312198 4967 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.312210 4967 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.312223 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.312235 4967 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.312247 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.312260 4967 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.312273 4967 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.312286 4967 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.312298 4967 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.312310 4967 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.312322 4967 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.312336 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.312349 4967 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.312361 4967 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.312374 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.312386 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.312398 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.312412 4967 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.312427 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.312439 4967 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.312452 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.312464 4967 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.312476 4967 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.312488 4967 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.312500 4967 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.312512 4967 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.312523 4967 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.312536 4967 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.312548 4967 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.312560 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.312572 4967 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.312584 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.312597 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.312610 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.312621 4967 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.312633 4967 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.312644 4967 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.312656 4967 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.312684 4967 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.312699 4967 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.312711 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.312722 4967 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.312731 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.312735 4967 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.312847 4967 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.312863 4967 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.312878 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.312893 4967 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.312907 4967 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.312921 4967 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.312936 4967 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.312948 4967 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.312959 4967 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.312971 4967 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.312983 4967 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.312997 4967 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.313010 4967 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.313024 4967 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.313037 4967 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.313049 4967 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.313062 4967 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.313075 4967 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.313089 4967 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.313103 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.313117 4967 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.313130 4967 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.313132 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.313141 4967 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.313155 4967 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.313170 4967 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.313184 4967 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.313199 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.313213 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.313227 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.313239 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.313252 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.313264 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.313277 4967 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.313289 4967 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.313302 4967 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.313315 4967 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.313328 4967 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.313342 4967 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.313354 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.313367 4967 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.313380 4967 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.313394 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.313408 4967 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.313421 4967 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.313434 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.313446 4967 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.313459 4967 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.313471 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.313485 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.313498 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.313510 4967 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.313523 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.313536 4967 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.313548 4967 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.313559 4967 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.313572 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.313586 4967 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.313598 4967 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.313612 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.313627 4967 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.313638 4967 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.313650 4967 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.313663 4967 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.313692 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.313704 4967 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.313716 4967 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.313729 4967 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.313743 4967 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.313755 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.313770 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.313783 4967 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.313797 4967 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.313811 4967 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.313826 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.313841 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.313855 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.313867 4967 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.313879 4967 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.313893 4967 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.313906 4967 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.313919 4967 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.313933 4967 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.313946 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.313959 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.313972 4967 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.313985 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.314000 4967 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.314013 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.314026 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.314038 4967 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.314050 4967 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.314062 4967 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.314075 4967 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.314089 4967 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.314102 4967 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.314115 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.314127 4967 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.314140 4967 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.314160 4967 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.314172 4967 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.314185 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.314179 4967 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="1400e110997c64fa33d6b8672c7e91cab5f782952ddf8e1dea2a882689026e7f" exitCode=255 Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.314200 4967 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.314213 4967 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.314226 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.314238 4967 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.314251 4967 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.314264 4967 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.314278 4967 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.314292 4967 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.314305 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.314319 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.314333 4967 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.314345 4967 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.314358 4967 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.314370 4967 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.314382 4967 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.314395 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.314407 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.314420 4967 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.314432 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.314444 4967 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.314457 4967 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.314470 4967 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.314483 4967 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.314495 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.314507 4967 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.314520 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.314533 4967 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.314545 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.314557 4967 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.314607 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"1400e110997c64fa33d6b8672c7e91cab5f782952ddf8e1dea2a882689026e7f"} Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.314683 4967 scope.go:117] "RemoveContainer" containerID="2c3118c3b932a91a617645840254d372bca2799d4b08eb2b371f5d2a0a6e116d" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.319847 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: E1008 00:07:03.321393 4967 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-crc\" already exists" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.324487 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.327332 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.329104 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.332416 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.332608 4967 scope.go:117] "RemoveContainer" containerID="1400e110997c64fa33d6b8672c7e91cab5f782952ddf8e1dea2a882689026e7f" Oct 08 00:07:03 crc kubenswrapper[4967]: E1008 00:07:03.332873 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.342479 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.344535 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.356218 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.365551 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58c4704-a88a-43b4-b944-4de626f13f7e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22f5dabb39da0a7bd01db4dae1437b9a231a1fcc0dfe749a4a5a637a72cf33c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01a2b3dc785520c79eae998482e5a1347725c76584357073238fe60a619edb77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6aa6f5e81feceab3d1c7587152a70546b365e629a69267e4d3084324a1de764a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63c2b7c4b2791bd1f16888c58b878dbff1f87884e68c9949de914d0d3e1d39fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.377519 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.377586 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 00:07:03 crc kubenswrapper[4967]: W1008 00:07:03.389646 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-3e45d0519ce26380610ee445ccb7935b070ef00cccbd613b61ceeb37f6f64f77 WatchSource:0}: Error finding container 3e45d0519ce26380610ee445ccb7935b070ef00cccbd613b61ceeb37f6f64f77: Status 404 returned error can't find the container with id 3e45d0519ce26380610ee445ccb7935b070ef00cccbd613b61ceeb37f6f64f77 Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.394561 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.395656 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.405169 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.415266 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.415491 4967 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.415515 4967 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.415525 4967 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.415534 4967 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 08 00:07:03 crc kubenswrapper[4967]: W1008 00:07:03.431786 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-3e2c8abcc12ab0fb29516cc8b3185eb12a932415a52be7721717bbd3536331ad WatchSource:0}: Error finding container 3e2c8abcc12ab0fb29516cc8b3185eb12a932415a52be7721717bbd3536331ad: Status 404 returned error can't find the container with id 3e2c8abcc12ab0fb29516cc8b3185eb12a932415a52be7721717bbd3536331ad Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.818310 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.818368 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.818393 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.818413 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 00:07:03 crc kubenswrapper[4967]: E1008 00:07:03.818467 4967 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 00:07:03 crc kubenswrapper[4967]: E1008 00:07:03.818573 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 00:07:03 crc kubenswrapper[4967]: E1008 00:07:03.818589 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 00:07:03 crc kubenswrapper[4967]: E1008 00:07:03.818600 4967 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 00:07:03 crc kubenswrapper[4967]: E1008 00:07:03.818651 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 00:07:03 crc kubenswrapper[4967]: E1008 00:07:03.818660 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 00:07:03 crc kubenswrapper[4967]: E1008 00:07:03.818685 4967 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 00:07:03 crc kubenswrapper[4967]: E1008 00:07:03.818466 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 00:07:04.818441883 +0000 UTC m=+23.468976532 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:07:03 crc kubenswrapper[4967]: I1008 00:07:03.818715 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 00:07:03 crc kubenswrapper[4967]: E1008 00:07:03.818741 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 00:07:04.818734041 +0000 UTC m=+23.469268690 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 00:07:03 crc kubenswrapper[4967]: E1008 00:07:03.818750 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-08 00:07:04.818745751 +0000 UTC m=+23.469280400 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 00:07:03 crc kubenswrapper[4967]: E1008 00:07:03.818759 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-08 00:07:04.818754811 +0000 UTC m=+23.469289460 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 00:07:03 crc kubenswrapper[4967]: E1008 00:07:03.818902 4967 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 00:07:03 crc kubenswrapper[4967]: E1008 00:07:03.818996 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 00:07:04.818976737 +0000 UTC m=+23.469511386 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.038575 4967 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.117155 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 00:07:04 crc kubenswrapper[4967]: E1008 00:07:04.117293 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.120864 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.121955 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.123969 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.124860 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.125912 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.126455 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.127156 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.128116 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.128698 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.129557 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.130078 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.131091 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.131537 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.132067 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.133010 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.133489 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.134839 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.135238 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.135802 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.136777 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.137224 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.138157 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.138629 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.139777 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.140285 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.140877 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.142017 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.142542 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.143520 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.144035 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.145048 4967 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.145166 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.146798 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.147650 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.148124 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.149941 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.150564 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.151471 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.152147 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.153189 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.153770 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.154944 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.155872 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.157054 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.157590 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.158475 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.159068 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.160429 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.161075 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.162202 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.162785 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.163945 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.164712 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.165222 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.305337 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-npz2s"] Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.305591 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-9tdbb"] Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.305753 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-npz2s" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.305764 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-9mrr7"] Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.305848 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-9tdbb" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.307355 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-xvdtp"] Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.309121 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-9mrr7" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.309451 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.310842 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.311414 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.312158 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.312256 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.312330 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.312379 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.312498 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.313419 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.313545 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.313717 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.313770 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.313899 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.314054 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.316524 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.317390 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.328594 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"94e63bc5641ce80d507f76a19b203f78c42368a7a4105f3fd09e19af21a1b735"} Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.328638 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"3e45d0519ce26380610ee445ccb7935b070ef00cccbd613b61ceeb37f6f64f77"} Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.330489 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.330822 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:04Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.334465 4967 scope.go:117] "RemoveContainer" containerID="1400e110997c64fa33d6b8672c7e91cab5f782952ddf8e1dea2a882689026e7f" Oct 08 00:07:04 crc kubenswrapper[4967]: E1008 00:07:04.334609 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.339119 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"3e2c8abcc12ab0fb29516cc8b3185eb12a932415a52be7721717bbd3536331ad"} Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.343001 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"5e98666fd3152bab57d3723e7da72aae2703b79f7ca73eb3c3e5f0e062a82592"} Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.343062 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"6cf1cbe5548146fe359aafa351864125ec6b8bdd7b796aa4d311f6f3a16ad212"} Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.343077 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"44b7a6a24f30155f39849227efadbb69b5e261413b58b26b972afd0028b1adc6"} Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.347660 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-npz2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c71501b1-aa9e-420e-b1bd-7e15365e7adc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gblk2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-npz2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:04Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.365260 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fa4b747-ab6e-4a11-9903-ef2dcb9668c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbadcdade274b6d9be940f22bc949e3c55cadf95dcad43c8cb7a1085c0e6708c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d0a0b2ef4a98ee828688081a74f4b7ecc31c8fae415d7257a560ecd43960ce3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96b53b1b283025b3d078a3f505d469d1872ad4bb8ab8e612c36c70b0b00a920\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1400e110997c64fa33d6b8672c7e91cab5f782952ddf8e1dea2a882689026e7f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c3118c3b932a91a617645840254d372bca2799d4b08eb2b371f5d2a0a6e116d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T00:06:46Z\\\",\\\"message\\\":\\\"W1008 00:06:45.407048 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 00:06:45.407307 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759882005 cert, and key in /tmp/serving-cert-1965353/serving-signer.crt, /tmp/serving-cert-1965353/serving-signer.key\\\\nI1008 00:06:45.750133 1 observer_polling.go:159] Starting file observer\\\\nW1008 00:06:45.755660 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 00:06:45.755870 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 00:06:45.756964 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1965353/tls.crt::/tmp/serving-cert-1965353/tls.key\\\\\\\"\\\\nF1008 00:06:46.020756 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1400e110997c64fa33d6b8672c7e91cab5f782952ddf8e1dea2a882689026e7f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 00:06:56.760088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 00:06:56.761210 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-369805205/tls.crt::/tmp/serving-cert-369805205/tls.key\\\\\\\"\\\\nI1008 00:07:02.321448 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 00:07:02.325932 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 00:07:02.325980 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 00:07:02.326025 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 00:07:02.326042 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 00:07:02.336203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 00:07:02.336240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 00:07:02.336245 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 00:07:02.336250 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 00:07:02.336253 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 00:07:02.336256 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 00:07:02.336258 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 00:07:02.336430 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 00:07:02.340774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca28d8db19036e47a7ef3dadb70941f6c6cf5735e7703977d036185c46fce40\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5b222ad84c31d8746dfe03124c7287b401deb3ca2c62ce2b4fbd40e638cee0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5b222ad84c31d8746dfe03124c7287b401deb3ca2c62ce2b4fbd40e638cee0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:04Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.378289 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:04Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.394353 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:04Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.407224 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58c4704-a88a-43b4-b944-4de626f13f7e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22f5dabb39da0a7bd01db4dae1437b9a231a1fcc0dfe749a4a5a637a72cf33c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01a2b3dc785520c79eae998482e5a1347725c76584357073238fe60a619edb77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6aa6f5e81feceab3d1c7587152a70546b365e629a69267e4d3084324a1de764a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63c2b7c4b2791bd1f16888c58b878dbff1f87884e68c9949de914d0d3e1d39fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:04Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.419317 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:04Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.425278 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/9461162a-c27d-465e-9ca7-2d3936877cd8-mcd-auth-proxy-config\") pod \"machine-config-daemon-xvdtp\" (UID: \"9461162a-c27d-465e-9ca7-2d3936877cd8\") " pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.425328 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/c71501b1-aa9e-420e-b1bd-7e15365e7adc-multus-socket-dir-parent\") pod \"multus-npz2s\" (UID: \"c71501b1-aa9e-420e-b1bd-7e15365e7adc\") " pod="openshift-multus/multus-npz2s" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.425353 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/41264a12-9486-4fee-a08c-607b590e5a70-hosts-file\") pod \"node-resolver-9tdbb\" (UID: \"41264a12-9486-4fee-a08c-607b590e5a70\") " pod="openshift-dns/node-resolver-9tdbb" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.425390 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/9461162a-c27d-465e-9ca7-2d3936877cd8-proxy-tls\") pod \"machine-config-daemon-xvdtp\" (UID: \"9461162a-c27d-465e-9ca7-2d3936877cd8\") " pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.425413 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c71501b1-aa9e-420e-b1bd-7e15365e7adc-host-var-lib-cni-bin\") pod \"multus-npz2s\" (UID: \"c71501b1-aa9e-420e-b1bd-7e15365e7adc\") " pod="openshift-multus/multus-npz2s" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.425467 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/c71501b1-aa9e-420e-b1bd-7e15365e7adc-multus-cni-dir\") pod \"multus-npz2s\" (UID: \"c71501b1-aa9e-420e-b1bd-7e15365e7adc\") " pod="openshift-multus/multus-npz2s" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.425514 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/9299bc4f-f8d3-4baf-9885-39573982fd64-cni-binary-copy\") pod \"multus-additional-cni-plugins-9mrr7\" (UID: \"9299bc4f-f8d3-4baf-9885-39573982fd64\") " pod="openshift-multus/multus-additional-cni-plugins-9mrr7" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.425543 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/c71501b1-aa9e-420e-b1bd-7e15365e7adc-multus-conf-dir\") pod \"multus-npz2s\" (UID: \"c71501b1-aa9e-420e-b1bd-7e15365e7adc\") " pod="openshift-multus/multus-npz2s" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.425592 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/9299bc4f-f8d3-4baf-9885-39573982fd64-os-release\") pod \"multus-additional-cni-plugins-9mrr7\" (UID: \"9299bc4f-f8d3-4baf-9885-39573982fd64\") " pod="openshift-multus/multus-additional-cni-plugins-9mrr7" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.425617 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c71501b1-aa9e-420e-b1bd-7e15365e7adc-etc-kubernetes\") pod \"multus-npz2s\" (UID: \"c71501b1-aa9e-420e-b1bd-7e15365e7adc\") " pod="openshift-multus/multus-npz2s" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.425651 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9299bc4f-f8d3-4baf-9885-39573982fd64-system-cni-dir\") pod \"multus-additional-cni-plugins-9mrr7\" (UID: \"9299bc4f-f8d3-4baf-9885-39573982fd64\") " pod="openshift-multus/multus-additional-cni-plugins-9mrr7" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.425689 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/c71501b1-aa9e-420e-b1bd-7e15365e7adc-cni-binary-copy\") pod \"multus-npz2s\" (UID: \"c71501b1-aa9e-420e-b1bd-7e15365e7adc\") " pod="openshift-multus/multus-npz2s" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.425711 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/c71501b1-aa9e-420e-b1bd-7e15365e7adc-host-var-lib-kubelet\") pod \"multus-npz2s\" (UID: \"c71501b1-aa9e-420e-b1bd-7e15365e7adc\") " pod="openshift-multus/multus-npz2s" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.425732 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/c71501b1-aa9e-420e-b1bd-7e15365e7adc-multus-daemon-config\") pod \"multus-npz2s\" (UID: \"c71501b1-aa9e-420e-b1bd-7e15365e7adc\") " pod="openshift-multus/multus-npz2s" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.425755 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c71501b1-aa9e-420e-b1bd-7e15365e7adc-host-run-netns\") pod \"multus-npz2s\" (UID: \"c71501b1-aa9e-420e-b1bd-7e15365e7adc\") " pod="openshift-multus/multus-npz2s" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.425776 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/c71501b1-aa9e-420e-b1bd-7e15365e7adc-host-var-lib-cni-multus\") pod \"multus-npz2s\" (UID: \"c71501b1-aa9e-420e-b1bd-7e15365e7adc\") " pod="openshift-multus/multus-npz2s" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.425797 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/c71501b1-aa9e-420e-b1bd-7e15365e7adc-host-run-multus-certs\") pod \"multus-npz2s\" (UID: \"c71501b1-aa9e-420e-b1bd-7e15365e7adc\") " pod="openshift-multus/multus-npz2s" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.425820 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/c71501b1-aa9e-420e-b1bd-7e15365e7adc-system-cni-dir\") pod \"multus-npz2s\" (UID: \"c71501b1-aa9e-420e-b1bd-7e15365e7adc\") " pod="openshift-multus/multus-npz2s" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.425894 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gblk2\" (UniqueName: \"kubernetes.io/projected/c71501b1-aa9e-420e-b1bd-7e15365e7adc-kube-api-access-gblk2\") pod \"multus-npz2s\" (UID: \"c71501b1-aa9e-420e-b1bd-7e15365e7adc\") " pod="openshift-multus/multus-npz2s" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.425930 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/9461162a-c27d-465e-9ca7-2d3936877cd8-rootfs\") pod \"machine-config-daemon-xvdtp\" (UID: \"9461162a-c27d-465e-9ca7-2d3936877cd8\") " pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.425972 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/9299bc4f-f8d3-4baf-9885-39573982fd64-cnibin\") pod \"multus-additional-cni-plugins-9mrr7\" (UID: \"9299bc4f-f8d3-4baf-9885-39573982fd64\") " pod="openshift-multus/multus-additional-cni-plugins-9mrr7" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.426024 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/9299bc4f-f8d3-4baf-9885-39573982fd64-tuning-conf-dir\") pod \"multus-additional-cni-plugins-9mrr7\" (UID: \"9299bc4f-f8d3-4baf-9885-39573982fd64\") " pod="openshift-multus/multus-additional-cni-plugins-9mrr7" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.426059 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/c71501b1-aa9e-420e-b1bd-7e15365e7adc-hostroot\") pod \"multus-npz2s\" (UID: \"c71501b1-aa9e-420e-b1bd-7e15365e7adc\") " pod="openshift-multus/multus-npz2s" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.426111 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s5blv\" (UniqueName: \"kubernetes.io/projected/41264a12-9486-4fee-a08c-607b590e5a70-kube-api-access-s5blv\") pod \"node-resolver-9tdbb\" (UID: \"41264a12-9486-4fee-a08c-607b590e5a70\") " pod="openshift-dns/node-resolver-9tdbb" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.426182 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z9tp4\" (UniqueName: \"kubernetes.io/projected/9461162a-c27d-465e-9ca7-2d3936877cd8-kube-api-access-z9tp4\") pod \"machine-config-daemon-xvdtp\" (UID: \"9461162a-c27d-465e-9ca7-2d3936877cd8\") " pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.426200 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dbsgx\" (UniqueName: \"kubernetes.io/projected/9299bc4f-f8d3-4baf-9885-39573982fd64-kube-api-access-dbsgx\") pod \"multus-additional-cni-plugins-9mrr7\" (UID: \"9299bc4f-f8d3-4baf-9885-39573982fd64\") " pod="openshift-multus/multus-additional-cni-plugins-9mrr7" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.426231 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/c71501b1-aa9e-420e-b1bd-7e15365e7adc-os-release\") pod \"multus-npz2s\" (UID: \"c71501b1-aa9e-420e-b1bd-7e15365e7adc\") " pod="openshift-multus/multus-npz2s" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.426249 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/c71501b1-aa9e-420e-b1bd-7e15365e7adc-host-run-k8s-cni-cncf-io\") pod \"multus-npz2s\" (UID: \"c71501b1-aa9e-420e-b1bd-7e15365e7adc\") " pod="openshift-multus/multus-npz2s" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.426277 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/9299bc4f-f8d3-4baf-9885-39573982fd64-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-9mrr7\" (UID: \"9299bc4f-f8d3-4baf-9885-39573982fd64\") " pod="openshift-multus/multus-additional-cni-plugins-9mrr7" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.426291 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/c71501b1-aa9e-420e-b1bd-7e15365e7adc-cnibin\") pod \"multus-npz2s\" (UID: \"c71501b1-aa9e-420e-b1bd-7e15365e7adc\") " pod="openshift-multus/multus-npz2s" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.433407 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:04Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.443976 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:04Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.459587 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:04Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.473822 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-npz2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c71501b1-aa9e-420e-b1bd-7e15365e7adc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gblk2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-npz2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:04Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.494944 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58c4704-a88a-43b4-b944-4de626f13f7e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22f5dabb39da0a7bd01db4dae1437b9a231a1fcc0dfe749a4a5a637a72cf33c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01a2b3dc785520c79eae998482e5a1347725c76584357073238fe60a619edb77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6aa6f5e81feceab3d1c7587152a70546b365e629a69267e4d3084324a1de764a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63c2b7c4b2791bd1f16888c58b878dbff1f87884e68c9949de914d0d3e1d39fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:04Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.511721 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:04Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.527300 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/9461162a-c27d-465e-9ca7-2d3936877cd8-rootfs\") pod \"machine-config-daemon-xvdtp\" (UID: \"9461162a-c27d-465e-9ca7-2d3936877cd8\") " pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.527353 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/9299bc4f-f8d3-4baf-9885-39573982fd64-cnibin\") pod \"multus-additional-cni-plugins-9mrr7\" (UID: \"9299bc4f-f8d3-4baf-9885-39573982fd64\") " pod="openshift-multus/multus-additional-cni-plugins-9mrr7" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.527377 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/9299bc4f-f8d3-4baf-9885-39573982fd64-tuning-conf-dir\") pod \"multus-additional-cni-plugins-9mrr7\" (UID: \"9299bc4f-f8d3-4baf-9885-39573982fd64\") " pod="openshift-multus/multus-additional-cni-plugins-9mrr7" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.527410 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/c71501b1-aa9e-420e-b1bd-7e15365e7adc-hostroot\") pod \"multus-npz2s\" (UID: \"c71501b1-aa9e-420e-b1bd-7e15365e7adc\") " pod="openshift-multus/multus-npz2s" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.527432 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s5blv\" (UniqueName: \"kubernetes.io/projected/41264a12-9486-4fee-a08c-607b590e5a70-kube-api-access-s5blv\") pod \"node-resolver-9tdbb\" (UID: \"41264a12-9486-4fee-a08c-607b590e5a70\") " pod="openshift-dns/node-resolver-9tdbb" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.527436 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/9461162a-c27d-465e-9ca7-2d3936877cd8-rootfs\") pod \"machine-config-daemon-xvdtp\" (UID: \"9461162a-c27d-465e-9ca7-2d3936877cd8\") " pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.527456 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z9tp4\" (UniqueName: \"kubernetes.io/projected/9461162a-c27d-465e-9ca7-2d3936877cd8-kube-api-access-z9tp4\") pod \"machine-config-daemon-xvdtp\" (UID: \"9461162a-c27d-465e-9ca7-2d3936877cd8\") " pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.527536 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dbsgx\" (UniqueName: \"kubernetes.io/projected/9299bc4f-f8d3-4baf-9885-39573982fd64-kube-api-access-dbsgx\") pod \"multus-additional-cni-plugins-9mrr7\" (UID: \"9299bc4f-f8d3-4baf-9885-39573982fd64\") " pod="openshift-multus/multus-additional-cni-plugins-9mrr7" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.527451 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/9299bc4f-f8d3-4baf-9885-39573982fd64-cnibin\") pod \"multus-additional-cni-plugins-9mrr7\" (UID: \"9299bc4f-f8d3-4baf-9885-39573982fd64\") " pod="openshift-multus/multus-additional-cni-plugins-9mrr7" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.527576 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/c71501b1-aa9e-420e-b1bd-7e15365e7adc-host-run-k8s-cni-cncf-io\") pod \"multus-npz2s\" (UID: \"c71501b1-aa9e-420e-b1bd-7e15365e7adc\") " pod="openshift-multus/multus-npz2s" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.527631 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/c71501b1-aa9e-420e-b1bd-7e15365e7adc-host-run-k8s-cni-cncf-io\") pod \"multus-npz2s\" (UID: \"c71501b1-aa9e-420e-b1bd-7e15365e7adc\") " pod="openshift-multus/multus-npz2s" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.527640 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/9299bc4f-f8d3-4baf-9885-39573982fd64-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-9mrr7\" (UID: \"9299bc4f-f8d3-4baf-9885-39573982fd64\") " pod="openshift-multus/multus-additional-cni-plugins-9mrr7" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.527492 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/c71501b1-aa9e-420e-b1bd-7e15365e7adc-hostroot\") pod \"multus-npz2s\" (UID: \"c71501b1-aa9e-420e-b1bd-7e15365e7adc\") " pod="openshift-multus/multus-npz2s" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.527708 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/c71501b1-aa9e-420e-b1bd-7e15365e7adc-cnibin\") pod \"multus-npz2s\" (UID: \"c71501b1-aa9e-420e-b1bd-7e15365e7adc\") " pod="openshift-multus/multus-npz2s" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.527773 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/c71501b1-aa9e-420e-b1bd-7e15365e7adc-cnibin\") pod \"multus-npz2s\" (UID: \"c71501b1-aa9e-420e-b1bd-7e15365e7adc\") " pod="openshift-multus/multus-npz2s" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.527796 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/c71501b1-aa9e-420e-b1bd-7e15365e7adc-os-release\") pod \"multus-npz2s\" (UID: \"c71501b1-aa9e-420e-b1bd-7e15365e7adc\") " pod="openshift-multus/multus-npz2s" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.527839 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/9461162a-c27d-465e-9ca7-2d3936877cd8-mcd-auth-proxy-config\") pod \"machine-config-daemon-xvdtp\" (UID: \"9461162a-c27d-465e-9ca7-2d3936877cd8\") " pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.527867 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/c71501b1-aa9e-420e-b1bd-7e15365e7adc-multus-socket-dir-parent\") pod \"multus-npz2s\" (UID: \"c71501b1-aa9e-420e-b1bd-7e15365e7adc\") " pod="openshift-multus/multus-npz2s" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.527888 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/41264a12-9486-4fee-a08c-607b590e5a70-hosts-file\") pod \"node-resolver-9tdbb\" (UID: \"41264a12-9486-4fee-a08c-607b590e5a70\") " pod="openshift-dns/node-resolver-9tdbb" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.527929 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/9461162a-c27d-465e-9ca7-2d3936877cd8-proxy-tls\") pod \"machine-config-daemon-xvdtp\" (UID: \"9461162a-c27d-465e-9ca7-2d3936877cd8\") " pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.527965 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/c71501b1-aa9e-420e-b1bd-7e15365e7adc-multus-cni-dir\") pod \"multus-npz2s\" (UID: \"c71501b1-aa9e-420e-b1bd-7e15365e7adc\") " pod="openshift-multus/multus-npz2s" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.527987 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c71501b1-aa9e-420e-b1bd-7e15365e7adc-host-var-lib-cni-bin\") pod \"multus-npz2s\" (UID: \"c71501b1-aa9e-420e-b1bd-7e15365e7adc\") " pod="openshift-multus/multus-npz2s" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.528010 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/9299bc4f-f8d3-4baf-9885-39573982fd64-cni-binary-copy\") pod \"multus-additional-cni-plugins-9mrr7\" (UID: \"9299bc4f-f8d3-4baf-9885-39573982fd64\") " pod="openshift-multus/multus-additional-cni-plugins-9mrr7" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.528017 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/41264a12-9486-4fee-a08c-607b590e5a70-hosts-file\") pod \"node-resolver-9tdbb\" (UID: \"41264a12-9486-4fee-a08c-607b590e5a70\") " pod="openshift-dns/node-resolver-9tdbb" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.528031 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/c71501b1-aa9e-420e-b1bd-7e15365e7adc-multus-conf-dir\") pod \"multus-npz2s\" (UID: \"c71501b1-aa9e-420e-b1bd-7e15365e7adc\") " pod="openshift-multus/multus-npz2s" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.528131 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/c71501b1-aa9e-420e-b1bd-7e15365e7adc-os-release\") pod \"multus-npz2s\" (UID: \"c71501b1-aa9e-420e-b1bd-7e15365e7adc\") " pod="openshift-multus/multus-npz2s" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.528157 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c71501b1-aa9e-420e-b1bd-7e15365e7adc-host-var-lib-cni-bin\") pod \"multus-npz2s\" (UID: \"c71501b1-aa9e-420e-b1bd-7e15365e7adc\") " pod="openshift-multus/multus-npz2s" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.528198 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/9299bc4f-f8d3-4baf-9885-39573982fd64-os-release\") pod \"multus-additional-cni-plugins-9mrr7\" (UID: \"9299bc4f-f8d3-4baf-9885-39573982fd64\") " pod="openshift-multus/multus-additional-cni-plugins-9mrr7" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.528200 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/c71501b1-aa9e-420e-b1bd-7e15365e7adc-multus-conf-dir\") pod \"multus-npz2s\" (UID: \"c71501b1-aa9e-420e-b1bd-7e15365e7adc\") " pod="openshift-multus/multus-npz2s" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.528218 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9299bc4f-f8d3-4baf-9885-39573982fd64-system-cni-dir\") pod \"multus-additional-cni-plugins-9mrr7\" (UID: \"9299bc4f-f8d3-4baf-9885-39573982fd64\") " pod="openshift-multus/multus-additional-cni-plugins-9mrr7" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.528199 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/c71501b1-aa9e-420e-b1bd-7e15365e7adc-multus-socket-dir-parent\") pod \"multus-npz2s\" (UID: \"c71501b1-aa9e-420e-b1bd-7e15365e7adc\") " pod="openshift-multus/multus-npz2s" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.528266 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/9299bc4f-f8d3-4baf-9885-39573982fd64-os-release\") pod \"multus-additional-cni-plugins-9mrr7\" (UID: \"9299bc4f-f8d3-4baf-9885-39573982fd64\") " pod="openshift-multus/multus-additional-cni-plugins-9mrr7" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.528276 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/9299bc4f-f8d3-4baf-9885-39573982fd64-tuning-conf-dir\") pod \"multus-additional-cni-plugins-9mrr7\" (UID: \"9299bc4f-f8d3-4baf-9885-39573982fd64\") " pod="openshift-multus/multus-additional-cni-plugins-9mrr7" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.528299 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9299bc4f-f8d3-4baf-9885-39573982fd64-system-cni-dir\") pod \"multus-additional-cni-plugins-9mrr7\" (UID: \"9299bc4f-f8d3-4baf-9885-39573982fd64\") " pod="openshift-multus/multus-additional-cni-plugins-9mrr7" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.528288 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/c71501b1-aa9e-420e-b1bd-7e15365e7adc-cni-binary-copy\") pod \"multus-npz2s\" (UID: \"c71501b1-aa9e-420e-b1bd-7e15365e7adc\") " pod="openshift-multus/multus-npz2s" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.528360 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/c71501b1-aa9e-420e-b1bd-7e15365e7adc-host-var-lib-kubelet\") pod \"multus-npz2s\" (UID: \"c71501b1-aa9e-420e-b1bd-7e15365e7adc\") " pod="openshift-multus/multus-npz2s" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.528312 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/c71501b1-aa9e-420e-b1bd-7e15365e7adc-multus-cni-dir\") pod \"multus-npz2s\" (UID: \"c71501b1-aa9e-420e-b1bd-7e15365e7adc\") " pod="openshift-multus/multus-npz2s" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.528394 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c71501b1-aa9e-420e-b1bd-7e15365e7adc-etc-kubernetes\") pod \"multus-npz2s\" (UID: \"c71501b1-aa9e-420e-b1bd-7e15365e7adc\") " pod="openshift-multus/multus-npz2s" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.528416 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/c71501b1-aa9e-420e-b1bd-7e15365e7adc-multus-daemon-config\") pod \"multus-npz2s\" (UID: \"c71501b1-aa9e-420e-b1bd-7e15365e7adc\") " pod="openshift-multus/multus-npz2s" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.528432 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c71501b1-aa9e-420e-b1bd-7e15365e7adc-host-run-netns\") pod \"multus-npz2s\" (UID: \"c71501b1-aa9e-420e-b1bd-7e15365e7adc\") " pod="openshift-multus/multus-npz2s" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.528451 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/c71501b1-aa9e-420e-b1bd-7e15365e7adc-host-var-lib-cni-multus\") pod \"multus-npz2s\" (UID: \"c71501b1-aa9e-420e-b1bd-7e15365e7adc\") " pod="openshift-multus/multus-npz2s" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.528455 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c71501b1-aa9e-420e-b1bd-7e15365e7adc-etc-kubernetes\") pod \"multus-npz2s\" (UID: \"c71501b1-aa9e-420e-b1bd-7e15365e7adc\") " pod="openshift-multus/multus-npz2s" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.528431 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/c71501b1-aa9e-420e-b1bd-7e15365e7adc-host-var-lib-kubelet\") pod \"multus-npz2s\" (UID: \"c71501b1-aa9e-420e-b1bd-7e15365e7adc\") " pod="openshift-multus/multus-npz2s" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.528473 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/c71501b1-aa9e-420e-b1bd-7e15365e7adc-host-run-multus-certs\") pod \"multus-npz2s\" (UID: \"c71501b1-aa9e-420e-b1bd-7e15365e7adc\") " pod="openshift-multus/multus-npz2s" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.528506 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/c71501b1-aa9e-420e-b1bd-7e15365e7adc-system-cni-dir\") pod \"multus-npz2s\" (UID: \"c71501b1-aa9e-420e-b1bd-7e15365e7adc\") " pod="openshift-multus/multus-npz2s" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.528510 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/c71501b1-aa9e-420e-b1bd-7e15365e7adc-host-var-lib-cni-multus\") pod \"multus-npz2s\" (UID: \"c71501b1-aa9e-420e-b1bd-7e15365e7adc\") " pod="openshift-multus/multus-npz2s" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.528480 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c71501b1-aa9e-420e-b1bd-7e15365e7adc-host-run-netns\") pod \"multus-npz2s\" (UID: \"c71501b1-aa9e-420e-b1bd-7e15365e7adc\") " pod="openshift-multus/multus-npz2s" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.528525 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gblk2\" (UniqueName: \"kubernetes.io/projected/c71501b1-aa9e-420e-b1bd-7e15365e7adc-kube-api-access-gblk2\") pod \"multus-npz2s\" (UID: \"c71501b1-aa9e-420e-b1bd-7e15365e7adc\") " pod="openshift-multus/multus-npz2s" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.528589 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/c71501b1-aa9e-420e-b1bd-7e15365e7adc-host-run-multus-certs\") pod \"multus-npz2s\" (UID: \"c71501b1-aa9e-420e-b1bd-7e15365e7adc\") " pod="openshift-multus/multus-npz2s" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.528630 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/c71501b1-aa9e-420e-b1bd-7e15365e7adc-system-cni-dir\") pod \"multus-npz2s\" (UID: \"c71501b1-aa9e-420e-b1bd-7e15365e7adc\") " pod="openshift-multus/multus-npz2s" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.528641 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/9461162a-c27d-465e-9ca7-2d3936877cd8-mcd-auth-proxy-config\") pod \"machine-config-daemon-xvdtp\" (UID: \"9461162a-c27d-465e-9ca7-2d3936877cd8\") " pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.529034 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/c71501b1-aa9e-420e-b1bd-7e15365e7adc-cni-binary-copy\") pod \"multus-npz2s\" (UID: \"c71501b1-aa9e-420e-b1bd-7e15365e7adc\") " pod="openshift-multus/multus-npz2s" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.529038 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/c71501b1-aa9e-420e-b1bd-7e15365e7adc-multus-daemon-config\") pod \"multus-npz2s\" (UID: \"c71501b1-aa9e-420e-b1bd-7e15365e7adc\") " pod="openshift-multus/multus-npz2s" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.529140 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/9299bc4f-f8d3-4baf-9885-39573982fd64-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-9mrr7\" (UID: \"9299bc4f-f8d3-4baf-9885-39573982fd64\") " pod="openshift-multus/multus-additional-cni-plugins-9mrr7" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.529174 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/9299bc4f-f8d3-4baf-9885-39573982fd64-cni-binary-copy\") pod \"multus-additional-cni-plugins-9mrr7\" (UID: \"9299bc4f-f8d3-4baf-9885-39573982fd64\") " pod="openshift-multus/multus-additional-cni-plugins-9mrr7" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.533641 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9mrr7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9299bc4f-f8d3-4baf-9885-39573982fd64\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9mrr7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:04Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.556976 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fa4b747-ab6e-4a11-9903-ef2dcb9668c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbadcdade274b6d9be940f22bc949e3c55cadf95dcad43c8cb7a1085c0e6708c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d0a0b2ef4a98ee828688081a74f4b7ecc31c8fae415d7257a560ecd43960ce3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96b53b1b283025b3d078a3f505d469d1872ad4bb8ab8e612c36c70b0b00a920\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1400e110997c64fa33d6b8672c7e91cab5f782952ddf8e1dea2a882689026e7f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1400e110997c64fa33d6b8672c7e91cab5f782952ddf8e1dea2a882689026e7f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 00:06:56.760088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 00:06:56.761210 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-369805205/tls.crt::/tmp/serving-cert-369805205/tls.key\\\\\\\"\\\\nI1008 00:07:02.321448 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 00:07:02.325932 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 00:07:02.325980 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 00:07:02.326025 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 00:07:02.326042 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 00:07:02.336203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 00:07:02.336240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 00:07:02.336245 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 00:07:02.336250 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 00:07:02.336253 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 00:07:02.336256 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 00:07:02.336258 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 00:07:02.336430 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 00:07:02.340774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:46Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca28d8db19036e47a7ef3dadb70941f6c6cf5735e7703977d036185c46fce40\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5b222ad84c31d8746dfe03124c7287b401deb3ca2c62ce2b4fbd40e638cee0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5b222ad84c31d8746dfe03124c7287b401deb3ca2c62ce2b4fbd40e638cee0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:04Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.578206 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:04Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.590862 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:04Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.605798 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9461162a-c27d-465e-9ca7-2d3936877cd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z9tp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z9tp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xvdtp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:04Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.606961 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/9461162a-c27d-465e-9ca7-2d3936877cd8-proxy-tls\") pod \"machine-config-daemon-xvdtp\" (UID: \"9461162a-c27d-465e-9ca7-2d3936877cd8\") " pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.607407 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s5blv\" (UniqueName: \"kubernetes.io/projected/41264a12-9486-4fee-a08c-607b590e5a70-kube-api-access-s5blv\") pod \"node-resolver-9tdbb\" (UID: \"41264a12-9486-4fee-a08c-607b590e5a70\") " pod="openshift-dns/node-resolver-9tdbb" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.607587 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z9tp4\" (UniqueName: \"kubernetes.io/projected/9461162a-c27d-465e-9ca7-2d3936877cd8-kube-api-access-z9tp4\") pod \"machine-config-daemon-xvdtp\" (UID: \"9461162a-c27d-465e-9ca7-2d3936877cd8\") " pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.623454 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94e63bc5641ce80d507f76a19b203f78c42368a7a4105f3fd09e19af21a1b735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:04Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.634111 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-9tdbb" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.650467 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e98666fd3152bab57d3723e7da72aae2703b79f7ca73eb3c3e5f0e062a82592\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cf1cbe5548146fe359aafa351864125ec6b8bdd7b796aa4d311f6f3a16ad212\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:04Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.654877 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.684442 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9tdbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41264a12-9486-4fee-a08c-607b590e5a70\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5blv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9tdbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:04Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.686833 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-bs8w9"] Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.687781 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.692787 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.692809 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.693065 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.693294 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.693557 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.693931 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.694059 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.707823 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gblk2\" (UniqueName: \"kubernetes.io/projected/c71501b1-aa9e-420e-b1bd-7e15365e7adc-kube-api-access-gblk2\") pod \"multus-npz2s\" (UID: \"c71501b1-aa9e-420e-b1bd-7e15365e7adc\") " pod="openshift-multus/multus-npz2s" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.707974 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dbsgx\" (UniqueName: \"kubernetes.io/projected/9299bc4f-f8d3-4baf-9885-39573982fd64-kube-api-access-dbsgx\") pod \"multus-additional-cni-plugins-9mrr7\" (UID: \"9299bc4f-f8d3-4baf-9885-39573982fd64\") " pod="openshift-multus/multus-additional-cni-plugins-9mrr7" Oct 08 00:07:04 crc kubenswrapper[4967]: W1008 00:07:04.712902 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9461162a_c27d_465e_9ca7_2d3936877cd8.slice/crio-600580c3bbcc648ef9e30148abb6ce2abd20dc692b24a8c048e18cb32b617857 WatchSource:0}: Error finding container 600580c3bbcc648ef9e30148abb6ce2abd20dc692b24a8c048e18cb32b617857: Status 404 returned error can't find the container with id 600580c3bbcc648ef9e30148abb6ce2abd20dc692b24a8c048e18cb32b617857 Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.748773 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:04Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.773619 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9mrr7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9299bc4f-f8d3-4baf-9885-39573982fd64\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9mrr7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:04Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.789800 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58c4704-a88a-43b4-b944-4de626f13f7e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22f5dabb39da0a7bd01db4dae1437b9a231a1fcc0dfe749a4a5a637a72cf33c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01a2b3dc785520c79eae998482e5a1347725c76584357073238fe60a619edb77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6aa6f5e81feceab3d1c7587152a70546b365e629a69267e4d3084324a1de764a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63c2b7c4b2791bd1f16888c58b878dbff1f87884e68c9949de914d0d3e1d39fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:04Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.807350 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:04Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.820952 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:04Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.833068 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.833161 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-var-lib-openvswitch\") pod \"ovnkube-node-bs8w9\" (UID: \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.833181 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-etc-openvswitch\") pod \"ovnkube-node-bs8w9\" (UID: \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" Oct 08 00:07:04 crc kubenswrapper[4967]: E1008 00:07:04.833204 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 00:07:06.833186245 +0000 UTC m=+25.483720894 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.833229 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.833259 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/4f1af2ba-8907-4d14-afa1-ffde207c80f4-ovnkube-script-lib\") pod \"ovnkube-node-bs8w9\" (UID: \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.834072 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/4f1af2ba-8907-4d14-afa1-ffde207c80f4-env-overrides\") pod \"ovnkube-node-bs8w9\" (UID: \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" Oct 08 00:07:04 crc kubenswrapper[4967]: E1008 00:07:04.833402 4967 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.834097 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-run-systemd\") pod \"ovnkube-node-bs8w9\" (UID: \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.834136 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/4f1af2ba-8907-4d14-afa1-ffde207c80f4-ovn-node-metrics-cert\") pod \"ovnkube-node-bs8w9\" (UID: \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.834158 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 00:07:04 crc kubenswrapper[4967]: E1008 00:07:04.834175 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 00:07:06.8341664 +0000 UTC m=+25.484701049 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.834191 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-run-ovn\") pod \"ovnkube-node-bs8w9\" (UID: \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.834213 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-log-socket\") pod \"ovnkube-node-bs8w9\" (UID: \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" Oct 08 00:07:04 crc kubenswrapper[4967]: E1008 00:07:04.834239 4967 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.834280 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.834331 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-host-kubelet\") pod \"ovnkube-node-bs8w9\" (UID: \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.834356 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-host-slash\") pod \"ovnkube-node-bs8w9\" (UID: \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" Oct 08 00:07:04 crc kubenswrapper[4967]: E1008 00:07:04.834384 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 00:07:06.834353424 +0000 UTC m=+25.484888143 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 00:07:04 crc kubenswrapper[4967]: E1008 00:07:04.834406 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.834409 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-bs8w9\" (UID: \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.834435 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/4f1af2ba-8907-4d14-afa1-ffde207c80f4-ovnkube-config\") pod \"ovnkube-node-bs8w9\" (UID: \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" Oct 08 00:07:04 crc kubenswrapper[4967]: E1008 00:07:04.834419 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 00:07:04 crc kubenswrapper[4967]: E1008 00:07:04.834464 4967 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.834466 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 00:07:04 crc kubenswrapper[4967]: E1008 00:07:04.834493 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-08 00:07:06.834484518 +0000 UTC m=+25.485019167 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.834524 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-run-openvswitch\") pod \"ovnkube-node-bs8w9\" (UID: \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.834561 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-host-cni-netd\") pod \"ovnkube-node-bs8w9\" (UID: \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.834582 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qlrlf\" (UniqueName: \"kubernetes.io/projected/4f1af2ba-8907-4d14-afa1-ffde207c80f4-kube-api-access-qlrlf\") pod \"ovnkube-node-bs8w9\" (UID: \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.834632 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-host-run-netns\") pod \"ovnkube-node-bs8w9\" (UID: \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.834652 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-node-log\") pod \"ovnkube-node-bs8w9\" (UID: \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" Oct 08 00:07:04 crc kubenswrapper[4967]: E1008 00:07:04.834707 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 00:07:04 crc kubenswrapper[4967]: E1008 00:07:04.834753 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 00:07:04 crc kubenswrapper[4967]: E1008 00:07:04.834770 4967 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.834726 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-systemd-units\") pod \"ovnkube-node-bs8w9\" (UID: \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" Oct 08 00:07:04 crc kubenswrapper[4967]: E1008 00:07:04.834838 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-08 00:07:06.834816026 +0000 UTC m=+25.485350755 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.834860 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-host-cni-bin\") pod \"ovnkube-node-bs8w9\" (UID: \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.834884 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-host-run-ovn-kubernetes\") pod \"ovnkube-node-bs8w9\" (UID: \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.837661 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9461162a-c27d-465e-9ca7-2d3936877cd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z9tp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z9tp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xvdtp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:04Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.853419 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fa4b747-ab6e-4a11-9903-ef2dcb9668c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbadcdade274b6d9be940f22bc949e3c55cadf95dcad43c8cb7a1085c0e6708c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d0a0b2ef4a98ee828688081a74f4b7ecc31c8fae415d7257a560ecd43960ce3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96b53b1b283025b3d078a3f505d469d1872ad4bb8ab8e612c36c70b0b00a920\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1400e110997c64fa33d6b8672c7e91cab5f782952ddf8e1dea2a882689026e7f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1400e110997c64fa33d6b8672c7e91cab5f782952ddf8e1dea2a882689026e7f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 00:06:56.760088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 00:06:56.761210 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-369805205/tls.crt::/tmp/serving-cert-369805205/tls.key\\\\\\\"\\\\nI1008 00:07:02.321448 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 00:07:02.325932 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 00:07:02.325980 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 00:07:02.326025 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 00:07:02.326042 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 00:07:02.336203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 00:07:02.336240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 00:07:02.336245 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 00:07:02.336250 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 00:07:02.336253 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 00:07:02.336256 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 00:07:02.336258 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 00:07:02.336430 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 00:07:02.340774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:46Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca28d8db19036e47a7ef3dadb70941f6c6cf5735e7703977d036185c46fce40\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5b222ad84c31d8746dfe03124c7287b401deb3ca2c62ce2b4fbd40e638cee0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5b222ad84c31d8746dfe03124c7287b401deb3ca2c62ce2b4fbd40e638cee0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:04Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.866118 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9tdbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41264a12-9486-4fee-a08c-607b590e5a70\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5blv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9tdbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:04Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.886482 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f1af2ba-8907-4d14-afa1-ffde207c80f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bs8w9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:04Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.899100 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94e63bc5641ce80d507f76a19b203f78c42368a7a4105f3fd09e19af21a1b735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:04Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.911883 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e98666fd3152bab57d3723e7da72aae2703b79f7ca73eb3c3e5f0e062a82592\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cf1cbe5548146fe359aafa351864125ec6b8bdd7b796aa4d311f6f3a16ad212\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:04Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.925902 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-npz2s" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.928095 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-npz2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c71501b1-aa9e-420e-b1bd-7e15365e7adc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gblk2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-npz2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:04Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.935953 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/4f1af2ba-8907-4d14-afa1-ffde207c80f4-env-overrides\") pod \"ovnkube-node-bs8w9\" (UID: \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.935994 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/4f1af2ba-8907-4d14-afa1-ffde207c80f4-ovnkube-script-lib\") pod \"ovnkube-node-bs8w9\" (UID: \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.936018 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-run-systemd\") pod \"ovnkube-node-bs8w9\" (UID: \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.936052 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/4f1af2ba-8907-4d14-afa1-ffde207c80f4-ovn-node-metrics-cert\") pod \"ovnkube-node-bs8w9\" (UID: \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.936090 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-run-ovn\") pod \"ovnkube-node-bs8w9\" (UID: \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.936127 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-host-kubelet\") pod \"ovnkube-node-bs8w9\" (UID: \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.936148 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-host-slash\") pod \"ovnkube-node-bs8w9\" (UID: \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.936191 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-run-systemd\") pod \"ovnkube-node-bs8w9\" (UID: \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.936230 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-log-socket\") pod \"ovnkube-node-bs8w9\" (UID: \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.936275 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-log-socket\") pod \"ovnkube-node-bs8w9\" (UID: \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.936316 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-bs8w9\" (UID: \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.936345 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/4f1af2ba-8907-4d14-afa1-ffde207c80f4-ovnkube-config\") pod \"ovnkube-node-bs8w9\" (UID: \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.936372 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qlrlf\" (UniqueName: \"kubernetes.io/projected/4f1af2ba-8907-4d14-afa1-ffde207c80f4-kube-api-access-qlrlf\") pod \"ovnkube-node-bs8w9\" (UID: \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.936389 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-run-openvswitch\") pod \"ovnkube-node-bs8w9\" (UID: \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.936403 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-host-cni-netd\") pod \"ovnkube-node-bs8w9\" (UID: \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.936426 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-host-run-netns\") pod \"ovnkube-node-bs8w9\" (UID: \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.936443 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-node-log\") pod \"ovnkube-node-bs8w9\" (UID: \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.936463 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-systemd-units\") pod \"ovnkube-node-bs8w9\" (UID: \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.936483 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-host-cni-bin\") pod \"ovnkube-node-bs8w9\" (UID: \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.936505 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-host-run-ovn-kubernetes\") pod \"ovnkube-node-bs8w9\" (UID: \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.936536 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-var-lib-openvswitch\") pod \"ovnkube-node-bs8w9\" (UID: \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.936578 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-etc-openvswitch\") pod \"ovnkube-node-bs8w9\" (UID: \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.936648 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-etc-openvswitch\") pod \"ovnkube-node-bs8w9\" (UID: \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.936697 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-bs8w9\" (UID: \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.936746 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/4f1af2ba-8907-4d14-afa1-ffde207c80f4-env-overrides\") pod \"ovnkube-node-bs8w9\" (UID: \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.936846 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-node-log\") pod \"ovnkube-node-bs8w9\" (UID: \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.936953 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-run-ovn\") pod \"ovnkube-node-bs8w9\" (UID: \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.936989 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-host-kubelet\") pod \"ovnkube-node-bs8w9\" (UID: \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.937013 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-host-slash\") pod \"ovnkube-node-bs8w9\" (UID: \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.937040 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-host-cni-bin\") pod \"ovnkube-node-bs8w9\" (UID: \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.937061 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-systemd-units\") pod \"ovnkube-node-bs8w9\" (UID: \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.937081 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-host-run-ovn-kubernetes\") pod \"ovnkube-node-bs8w9\" (UID: \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.937106 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-var-lib-openvswitch\") pod \"ovnkube-node-bs8w9\" (UID: \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.937128 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-host-cni-netd\") pod \"ovnkube-node-bs8w9\" (UID: \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.937151 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-run-openvswitch\") pod \"ovnkube-node-bs8w9\" (UID: \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.937173 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-host-run-netns\") pod \"ovnkube-node-bs8w9\" (UID: \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.937434 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/4f1af2ba-8907-4d14-afa1-ffde207c80f4-ovnkube-config\") pod \"ovnkube-node-bs8w9\" (UID: \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.937506 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/4f1af2ba-8907-4d14-afa1-ffde207c80f4-ovnkube-script-lib\") pod \"ovnkube-node-bs8w9\" (UID: \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" Oct 08 00:07:04 crc kubenswrapper[4967]: W1008 00:07:04.942700 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc71501b1_aa9e_420e_b1bd_7e15365e7adc.slice/crio-f31dad52a91e729110fd7e9167673567842e939f0cb8d6aa99f133727c32cb2a WatchSource:0}: Error finding container f31dad52a91e729110fd7e9167673567842e939f0cb8d6aa99f133727c32cb2a: Status 404 returned error can't find the container with id f31dad52a91e729110fd7e9167673567842e939f0cb8d6aa99f133727c32cb2a Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.943039 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/4f1af2ba-8907-4d14-afa1-ffde207c80f4-ovn-node-metrics-cert\") pod \"ovnkube-node-bs8w9\" (UID: \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.945398 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-9mrr7" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.947720 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:04Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:04 crc kubenswrapper[4967]: I1008 00:07:04.962043 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qlrlf\" (UniqueName: \"kubernetes.io/projected/4f1af2ba-8907-4d14-afa1-ffde207c80f4-kube-api-access-qlrlf\") pod \"ovnkube-node-bs8w9\" (UID: \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\") " pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" Oct 08 00:07:05 crc kubenswrapper[4967]: I1008 00:07:05.052881 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" Oct 08 00:07:05 crc kubenswrapper[4967]: I1008 00:07:05.117812 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 00:07:05 crc kubenswrapper[4967]: E1008 00:07:05.117927 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 00:07:05 crc kubenswrapper[4967]: I1008 00:07:05.117988 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 00:07:05 crc kubenswrapper[4967]: E1008 00:07:05.118894 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 00:07:05 crc kubenswrapper[4967]: I1008 00:07:05.347458 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-npz2s" event={"ID":"c71501b1-aa9e-420e-b1bd-7e15365e7adc","Type":"ContainerStarted","Data":"82dfe6c420d804b2d7daef4b5cfe3f87b77b369287c4d65553fa93f5f24f589d"} Oct 08 00:07:05 crc kubenswrapper[4967]: I1008 00:07:05.347505 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-npz2s" event={"ID":"c71501b1-aa9e-420e-b1bd-7e15365e7adc","Type":"ContainerStarted","Data":"f31dad52a91e729110fd7e9167673567842e939f0cb8d6aa99f133727c32cb2a"} Oct 08 00:07:05 crc kubenswrapper[4967]: I1008 00:07:05.349942 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" event={"ID":"9461162a-c27d-465e-9ca7-2d3936877cd8","Type":"ContainerStarted","Data":"e9778d78ddc7325abd4d77bc5e71b0b114f50be5dda4a438c7cb6b05640c8a07"} Oct 08 00:07:05 crc kubenswrapper[4967]: I1008 00:07:05.349970 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" event={"ID":"9461162a-c27d-465e-9ca7-2d3936877cd8","Type":"ContainerStarted","Data":"28760af04ef321d0f91b2bce3ad074d44b63dabf676cc37f3be6aa8726341a2f"} Oct 08 00:07:05 crc kubenswrapper[4967]: I1008 00:07:05.349979 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" event={"ID":"9461162a-c27d-465e-9ca7-2d3936877cd8","Type":"ContainerStarted","Data":"600580c3bbcc648ef9e30148abb6ce2abd20dc692b24a8c048e18cb32b617857"} Oct 08 00:07:05 crc kubenswrapper[4967]: I1008 00:07:05.351372 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-9tdbb" event={"ID":"41264a12-9486-4fee-a08c-607b590e5a70","Type":"ContainerStarted","Data":"806e85df796ae300d279f73a0cc4433a5bb1f7620ee8ce33c8dc4674c4ce0596"} Oct 08 00:07:05 crc kubenswrapper[4967]: I1008 00:07:05.351401 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-9tdbb" event={"ID":"41264a12-9486-4fee-a08c-607b590e5a70","Type":"ContainerStarted","Data":"f29efea5ddf623e5ffd6778485ad8208c7a1f3785e971d24a7d7b8b09313b43c"} Oct 08 00:07:05 crc kubenswrapper[4967]: I1008 00:07:05.352687 4967 generic.go:334] "Generic (PLEG): container finished" podID="4f1af2ba-8907-4d14-afa1-ffde207c80f4" containerID="94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427" exitCode=0 Oct 08 00:07:05 crc kubenswrapper[4967]: I1008 00:07:05.352729 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" event={"ID":"4f1af2ba-8907-4d14-afa1-ffde207c80f4","Type":"ContainerDied","Data":"94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427"} Oct 08 00:07:05 crc kubenswrapper[4967]: I1008 00:07:05.352744 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" event={"ID":"4f1af2ba-8907-4d14-afa1-ffde207c80f4","Type":"ContainerStarted","Data":"11f2eddeaf743f73001e819048455fe3ce59f7dbad311bd1525a412bf99ff182"} Oct 08 00:07:05 crc kubenswrapper[4967]: I1008 00:07:05.354732 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9mrr7" event={"ID":"9299bc4f-f8d3-4baf-9885-39573982fd64","Type":"ContainerStarted","Data":"fb49c8476b482416588c4d403016940730064a619c0641eb029c947a4bab1215"} Oct 08 00:07:05 crc kubenswrapper[4967]: I1008 00:07:05.354754 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9mrr7" event={"ID":"9299bc4f-f8d3-4baf-9885-39573982fd64","Type":"ContainerStarted","Data":"1322634f658c777e81c11181084b903941d1723aaa1fee97bb1569457ea44bb5"} Oct 08 00:07:05 crc kubenswrapper[4967]: I1008 00:07:05.355141 4967 scope.go:117] "RemoveContainer" containerID="1400e110997c64fa33d6b8672c7e91cab5f782952ddf8e1dea2a882689026e7f" Oct 08 00:07:05 crc kubenswrapper[4967]: E1008 00:07:05.355250 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Oct 08 00:07:05 crc kubenswrapper[4967]: I1008 00:07:05.371343 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:05Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:05 crc kubenswrapper[4967]: I1008 00:07:05.407184 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9mrr7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9299bc4f-f8d3-4baf-9885-39573982fd64\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9mrr7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:05Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:05 crc kubenswrapper[4967]: I1008 00:07:05.433730 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58c4704-a88a-43b4-b944-4de626f13f7e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22f5dabb39da0a7bd01db4dae1437b9a231a1fcc0dfe749a4a5a637a72cf33c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01a2b3dc785520c79eae998482e5a1347725c76584357073238fe60a619edb77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6aa6f5e81feceab3d1c7587152a70546b365e629a69267e4d3084324a1de764a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63c2b7c4b2791bd1f16888c58b878dbff1f87884e68c9949de914d0d3e1d39fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:05Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:05 crc kubenswrapper[4967]: I1008 00:07:05.452649 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:05Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:05 crc kubenswrapper[4967]: I1008 00:07:05.468399 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:05Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:05 crc kubenswrapper[4967]: I1008 00:07:05.481457 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9461162a-c27d-465e-9ca7-2d3936877cd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z9tp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z9tp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xvdtp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:05Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:05 crc kubenswrapper[4967]: I1008 00:07:05.497044 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fa4b747-ab6e-4a11-9903-ef2dcb9668c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbadcdade274b6d9be940f22bc949e3c55cadf95dcad43c8cb7a1085c0e6708c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d0a0b2ef4a98ee828688081a74f4b7ecc31c8fae415d7257a560ecd43960ce3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96b53b1b283025b3d078a3f505d469d1872ad4bb8ab8e612c36c70b0b00a920\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1400e110997c64fa33d6b8672c7e91cab5f782952ddf8e1dea2a882689026e7f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1400e110997c64fa33d6b8672c7e91cab5f782952ddf8e1dea2a882689026e7f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 00:06:56.760088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 00:06:56.761210 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-369805205/tls.crt::/tmp/serving-cert-369805205/tls.key\\\\\\\"\\\\nI1008 00:07:02.321448 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 00:07:02.325932 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 00:07:02.325980 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 00:07:02.326025 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 00:07:02.326042 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 00:07:02.336203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 00:07:02.336240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 00:07:02.336245 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 00:07:02.336250 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 00:07:02.336253 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 00:07:02.336256 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 00:07:02.336258 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 00:07:02.336430 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 00:07:02.340774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:46Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca28d8db19036e47a7ef3dadb70941f6c6cf5735e7703977d036185c46fce40\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5b222ad84c31d8746dfe03124c7287b401deb3ca2c62ce2b4fbd40e638cee0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5b222ad84c31d8746dfe03124c7287b401deb3ca2c62ce2b4fbd40e638cee0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:05Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:05 crc kubenswrapper[4967]: I1008 00:07:05.511371 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9tdbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41264a12-9486-4fee-a08c-607b590e5a70\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5blv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9tdbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:05Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:05 crc kubenswrapper[4967]: I1008 00:07:05.530249 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f1af2ba-8907-4d14-afa1-ffde207c80f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bs8w9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:05Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:05 crc kubenswrapper[4967]: I1008 00:07:05.545021 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94e63bc5641ce80d507f76a19b203f78c42368a7a4105f3fd09e19af21a1b735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:05Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:05 crc kubenswrapper[4967]: I1008 00:07:05.561694 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e98666fd3152bab57d3723e7da72aae2703b79f7ca73eb3c3e5f0e062a82592\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cf1cbe5548146fe359aafa351864125ec6b8bdd7b796aa4d311f6f3a16ad212\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:05Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:05 crc kubenswrapper[4967]: I1008 00:07:05.573979 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-npz2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c71501b1-aa9e-420e-b1bd-7e15365e7adc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dfe6c420d804b2d7daef4b5cfe3f87b77b369287c4d65553fa93f5f24f589d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gblk2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-npz2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:05Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:05 crc kubenswrapper[4967]: I1008 00:07:05.585956 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:05Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:05 crc kubenswrapper[4967]: I1008 00:07:05.606897 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58c4704-a88a-43b4-b944-4de626f13f7e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22f5dabb39da0a7bd01db4dae1437b9a231a1fcc0dfe749a4a5a637a72cf33c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01a2b3dc785520c79eae998482e5a1347725c76584357073238fe60a619edb77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6aa6f5e81feceab3d1c7587152a70546b365e629a69267e4d3084324a1de764a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63c2b7c4b2791bd1f16888c58b878dbff1f87884e68c9949de914d0d3e1d39fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:05Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:05 crc kubenswrapper[4967]: I1008 00:07:05.619550 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:05Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:05 crc kubenswrapper[4967]: I1008 00:07:05.636403 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9mrr7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9299bc4f-f8d3-4baf-9885-39573982fd64\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb49c8476b482416588c4d403016940730064a619c0641eb029c947a4bab1215\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9mrr7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:05Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:05 crc kubenswrapper[4967]: I1008 00:07:05.645956 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9461162a-c27d-465e-9ca7-2d3936877cd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9778d78ddc7325abd4d77bc5e71b0b114f50be5dda4a438c7cb6b05640c8a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z9tp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28760af04ef321d0f91b2bce3ad074d44b63dabf676cc37f3be6aa8726341a2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z9tp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xvdtp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:05Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:05 crc kubenswrapper[4967]: I1008 00:07:05.667223 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fa4b747-ab6e-4a11-9903-ef2dcb9668c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbadcdade274b6d9be940f22bc949e3c55cadf95dcad43c8cb7a1085c0e6708c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d0a0b2ef4a98ee828688081a74f4b7ecc31c8fae415d7257a560ecd43960ce3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96b53b1b283025b3d078a3f505d469d1872ad4bb8ab8e612c36c70b0b00a920\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1400e110997c64fa33d6b8672c7e91cab5f782952ddf8e1dea2a882689026e7f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1400e110997c64fa33d6b8672c7e91cab5f782952ddf8e1dea2a882689026e7f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 00:06:56.760088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 00:06:56.761210 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-369805205/tls.crt::/tmp/serving-cert-369805205/tls.key\\\\\\\"\\\\nI1008 00:07:02.321448 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 00:07:02.325932 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 00:07:02.325980 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 00:07:02.326025 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 00:07:02.326042 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 00:07:02.336203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 00:07:02.336240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 00:07:02.336245 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 00:07:02.336250 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 00:07:02.336253 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 00:07:02.336256 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 00:07:02.336258 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 00:07:02.336430 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 00:07:02.340774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:46Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca28d8db19036e47a7ef3dadb70941f6c6cf5735e7703977d036185c46fce40\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5b222ad84c31d8746dfe03124c7287b401deb3ca2c62ce2b4fbd40e638cee0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5b222ad84c31d8746dfe03124c7287b401deb3ca2c62ce2b4fbd40e638cee0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:05Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:05 crc kubenswrapper[4967]: I1008 00:07:05.685322 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:05Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:05 crc kubenswrapper[4967]: I1008 00:07:05.696838 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:05Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:05 crc kubenswrapper[4967]: I1008 00:07:05.712196 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94e63bc5641ce80d507f76a19b203f78c42368a7a4105f3fd09e19af21a1b735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:05Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:05 crc kubenswrapper[4967]: I1008 00:07:05.726814 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e98666fd3152bab57d3723e7da72aae2703b79f7ca73eb3c3e5f0e062a82592\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cf1cbe5548146fe359aafa351864125ec6b8bdd7b796aa4d311f6f3a16ad212\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:05Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:05 crc kubenswrapper[4967]: I1008 00:07:05.736288 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9tdbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41264a12-9486-4fee-a08c-607b590e5a70\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://806e85df796ae300d279f73a0cc4433a5bb1f7620ee8ce33c8dc4674c4ce0596\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5blv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9tdbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:05Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:05 crc kubenswrapper[4967]: I1008 00:07:05.754357 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f1af2ba-8907-4d14-afa1-ffde207c80f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bs8w9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:05Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:05 crc kubenswrapper[4967]: I1008 00:07:05.773277 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:05Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:05 crc kubenswrapper[4967]: I1008 00:07:05.790363 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-npz2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c71501b1-aa9e-420e-b1bd-7e15365e7adc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dfe6c420d804b2d7daef4b5cfe3f87b77b369287c4d65553fa93f5f24f589d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gblk2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-npz2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:05Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:06 crc kubenswrapper[4967]: I1008 00:07:06.116915 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 00:07:06 crc kubenswrapper[4967]: E1008 00:07:06.117042 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 00:07:06 crc kubenswrapper[4967]: I1008 00:07:06.193008 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-v5n7t"] Oct 08 00:07:06 crc kubenswrapper[4967]: I1008 00:07:06.193304 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-v5n7t" Oct 08 00:07:06 crc kubenswrapper[4967]: I1008 00:07:06.196272 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Oct 08 00:07:06 crc kubenswrapper[4967]: I1008 00:07:06.196297 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Oct 08 00:07:06 crc kubenswrapper[4967]: I1008 00:07:06.196584 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Oct 08 00:07:06 crc kubenswrapper[4967]: I1008 00:07:06.196644 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Oct 08 00:07:06 crc kubenswrapper[4967]: I1008 00:07:06.208374 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:06Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:06 crc kubenswrapper[4967]: I1008 00:07:06.222569 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:06Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:06 crc kubenswrapper[4967]: I1008 00:07:06.234104 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9461162a-c27d-465e-9ca7-2d3936877cd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9778d78ddc7325abd4d77bc5e71b0b114f50be5dda4a438c7cb6b05640c8a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z9tp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28760af04ef321d0f91b2bce3ad074d44b63dabf676cc37f3be6aa8726341a2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z9tp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xvdtp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:06Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:06 crc kubenswrapper[4967]: I1008 00:07:06.247960 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fa4b747-ab6e-4a11-9903-ef2dcb9668c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbadcdade274b6d9be940f22bc949e3c55cadf95dcad43c8cb7a1085c0e6708c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d0a0b2ef4a98ee828688081a74f4b7ecc31c8fae415d7257a560ecd43960ce3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96b53b1b283025b3d078a3f505d469d1872ad4bb8ab8e612c36c70b0b00a920\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1400e110997c64fa33d6b8672c7e91cab5f782952ddf8e1dea2a882689026e7f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1400e110997c64fa33d6b8672c7e91cab5f782952ddf8e1dea2a882689026e7f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 00:06:56.760088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 00:06:56.761210 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-369805205/tls.crt::/tmp/serving-cert-369805205/tls.key\\\\\\\"\\\\nI1008 00:07:02.321448 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 00:07:02.325932 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 00:07:02.325980 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 00:07:02.326025 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 00:07:02.326042 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 00:07:02.336203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 00:07:02.336240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 00:07:02.336245 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 00:07:02.336250 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 00:07:02.336253 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 00:07:02.336256 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 00:07:02.336258 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 00:07:02.336430 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 00:07:02.340774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:46Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca28d8db19036e47a7ef3dadb70941f6c6cf5735e7703977d036185c46fce40\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5b222ad84c31d8746dfe03124c7287b401deb3ca2c62ce2b4fbd40e638cee0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5b222ad84c31d8746dfe03124c7287b401deb3ca2c62ce2b4fbd40e638cee0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:06Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:06 crc kubenswrapper[4967]: I1008 00:07:06.260900 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e98666fd3152bab57d3723e7da72aae2703b79f7ca73eb3c3e5f0e062a82592\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cf1cbe5548146fe359aafa351864125ec6b8bdd7b796aa4d311f6f3a16ad212\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:06Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:06 crc kubenswrapper[4967]: I1008 00:07:06.271803 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9tdbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41264a12-9486-4fee-a08c-607b590e5a70\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://806e85df796ae300d279f73a0cc4433a5bb1f7620ee8ce33c8dc4674c4ce0596\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5blv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9tdbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:06Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:06 crc kubenswrapper[4967]: I1008 00:07:06.285965 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bpsjd\" (UniqueName: \"kubernetes.io/projected/abe7ae2d-7e63-4249-96e1-4000e9628ccc-kube-api-access-bpsjd\") pod \"node-ca-v5n7t\" (UID: \"abe7ae2d-7e63-4249-96e1-4000e9628ccc\") " pod="openshift-image-registry/node-ca-v5n7t" Oct 08 00:07:06 crc kubenswrapper[4967]: I1008 00:07:06.286014 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/abe7ae2d-7e63-4249-96e1-4000e9628ccc-serviceca\") pod \"node-ca-v5n7t\" (UID: \"abe7ae2d-7e63-4249-96e1-4000e9628ccc\") " pod="openshift-image-registry/node-ca-v5n7t" Oct 08 00:07:06 crc kubenswrapper[4967]: I1008 00:07:06.286042 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/abe7ae2d-7e63-4249-96e1-4000e9628ccc-host\") pod \"node-ca-v5n7t\" (UID: \"abe7ae2d-7e63-4249-96e1-4000e9628ccc\") " pod="openshift-image-registry/node-ca-v5n7t" Oct 08 00:07:06 crc kubenswrapper[4967]: I1008 00:07:06.291720 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f1af2ba-8907-4d14-afa1-ffde207c80f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bs8w9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:06Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:06 crc kubenswrapper[4967]: I1008 00:07:06.302836 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-v5n7t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abe7ae2d-7e63-4249-96e1-4000e9628ccc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpsjd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:06Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-v5n7t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:06Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:06 crc kubenswrapper[4967]: I1008 00:07:06.314914 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94e63bc5641ce80d507f76a19b203f78c42368a7a4105f3fd09e19af21a1b735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:06Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:06 crc kubenswrapper[4967]: I1008 00:07:06.328595 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:06Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:06 crc kubenswrapper[4967]: I1008 00:07:06.341390 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-npz2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c71501b1-aa9e-420e-b1bd-7e15365e7adc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dfe6c420d804b2d7daef4b5cfe3f87b77b369287c4d65553fa93f5f24f589d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gblk2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-npz2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:06Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:06 crc kubenswrapper[4967]: I1008 00:07:06.355579 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58c4704-a88a-43b4-b944-4de626f13f7e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22f5dabb39da0a7bd01db4dae1437b9a231a1fcc0dfe749a4a5a637a72cf33c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01a2b3dc785520c79eae998482e5a1347725c76584357073238fe60a619edb77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6aa6f5e81feceab3d1c7587152a70546b365e629a69267e4d3084324a1de764a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63c2b7c4b2791bd1f16888c58b878dbff1f87884e68c9949de914d0d3e1d39fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:06Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:06 crc kubenswrapper[4967]: I1008 00:07:06.361880 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" event={"ID":"4f1af2ba-8907-4d14-afa1-ffde207c80f4","Type":"ContainerStarted","Data":"9c9fa0559eb35ba67e11e36069ac8d0b2fec76399788eef668a08ad4353df3ab"} Oct 08 00:07:06 crc kubenswrapper[4967]: I1008 00:07:06.361924 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" event={"ID":"4f1af2ba-8907-4d14-afa1-ffde207c80f4","Type":"ContainerStarted","Data":"d3312263e12bd42fbed0476ab5ace66b5497f62c9477865e130d67e3922e05cb"} Oct 08 00:07:06 crc kubenswrapper[4967]: I1008 00:07:06.361938 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" event={"ID":"4f1af2ba-8907-4d14-afa1-ffde207c80f4","Type":"ContainerStarted","Data":"b03b83ddff429dbfa8fe052901aaa86a39ff0dd75477ffa8a6b6f4198f5819ca"} Oct 08 00:07:06 crc kubenswrapper[4967]: I1008 00:07:06.361949 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" event={"ID":"4f1af2ba-8907-4d14-afa1-ffde207c80f4","Type":"ContainerStarted","Data":"4535cffb8c9d64e662e282ecb44f6817b4604866f0bb75015f3b888d5554f821"} Oct 08 00:07:06 crc kubenswrapper[4967]: I1008 00:07:06.361960 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" event={"ID":"4f1af2ba-8907-4d14-afa1-ffde207c80f4","Type":"ContainerStarted","Data":"6e69dc603055aa221d44d11c049ddcc50cec568979911a310ed773a90341a499"} Oct 08 00:07:06 crc kubenswrapper[4967]: I1008 00:07:06.361971 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" event={"ID":"4f1af2ba-8907-4d14-afa1-ffde207c80f4","Type":"ContainerStarted","Data":"115f7a33f3dedae8f59d2f45f465f95f48b99c824e895ef7a0be9fe7a1b61d23"} Oct 08 00:07:06 crc kubenswrapper[4967]: I1008 00:07:06.363847 4967 generic.go:334] "Generic (PLEG): container finished" podID="9299bc4f-f8d3-4baf-9885-39573982fd64" containerID="fb49c8476b482416588c4d403016940730064a619c0641eb029c947a4bab1215" exitCode=0 Oct 08 00:07:06 crc kubenswrapper[4967]: I1008 00:07:06.363900 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9mrr7" event={"ID":"9299bc4f-f8d3-4baf-9885-39573982fd64","Type":"ContainerDied","Data":"fb49c8476b482416588c4d403016940730064a619c0641eb029c947a4bab1215"} Oct 08 00:07:06 crc kubenswrapper[4967]: I1008 00:07:06.365368 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"dbfcdc882e2be1cb9cf666c9e7775671078f7dbbfe8421dd15db737b89920a36"} Oct 08 00:07:06 crc kubenswrapper[4967]: I1008 00:07:06.375896 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:06Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:06 crc kubenswrapper[4967]: I1008 00:07:06.386495 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bpsjd\" (UniqueName: \"kubernetes.io/projected/abe7ae2d-7e63-4249-96e1-4000e9628ccc-kube-api-access-bpsjd\") pod \"node-ca-v5n7t\" (UID: \"abe7ae2d-7e63-4249-96e1-4000e9628ccc\") " pod="openshift-image-registry/node-ca-v5n7t" Oct 08 00:07:06 crc kubenswrapper[4967]: I1008 00:07:06.386527 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/abe7ae2d-7e63-4249-96e1-4000e9628ccc-serviceca\") pod \"node-ca-v5n7t\" (UID: \"abe7ae2d-7e63-4249-96e1-4000e9628ccc\") " pod="openshift-image-registry/node-ca-v5n7t" Oct 08 00:07:06 crc kubenswrapper[4967]: I1008 00:07:06.386548 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/abe7ae2d-7e63-4249-96e1-4000e9628ccc-host\") pod \"node-ca-v5n7t\" (UID: \"abe7ae2d-7e63-4249-96e1-4000e9628ccc\") " pod="openshift-image-registry/node-ca-v5n7t" Oct 08 00:07:06 crc kubenswrapper[4967]: I1008 00:07:06.386602 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/abe7ae2d-7e63-4249-96e1-4000e9628ccc-host\") pod \"node-ca-v5n7t\" (UID: \"abe7ae2d-7e63-4249-96e1-4000e9628ccc\") " pod="openshift-image-registry/node-ca-v5n7t" Oct 08 00:07:06 crc kubenswrapper[4967]: I1008 00:07:06.391076 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/abe7ae2d-7e63-4249-96e1-4000e9628ccc-serviceca\") pod \"node-ca-v5n7t\" (UID: \"abe7ae2d-7e63-4249-96e1-4000e9628ccc\") " pod="openshift-image-registry/node-ca-v5n7t" Oct 08 00:07:06 crc kubenswrapper[4967]: I1008 00:07:06.401506 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9mrr7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9299bc4f-f8d3-4baf-9885-39573982fd64\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb49c8476b482416588c4d403016940730064a619c0641eb029c947a4bab1215\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9mrr7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:06Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:06 crc kubenswrapper[4967]: I1008 00:07:06.418486 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bpsjd\" (UniqueName: \"kubernetes.io/projected/abe7ae2d-7e63-4249-96e1-4000e9628ccc-kube-api-access-bpsjd\") pod \"node-ca-v5n7t\" (UID: \"abe7ae2d-7e63-4249-96e1-4000e9628ccc\") " pod="openshift-image-registry/node-ca-v5n7t" Oct 08 00:07:06 crc kubenswrapper[4967]: I1008 00:07:06.424132 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94e63bc5641ce80d507f76a19b203f78c42368a7a4105f3fd09e19af21a1b735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:06Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:06 crc kubenswrapper[4967]: I1008 00:07:06.438523 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e98666fd3152bab57d3723e7da72aae2703b79f7ca73eb3c3e5f0e062a82592\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cf1cbe5548146fe359aafa351864125ec6b8bdd7b796aa4d311f6f3a16ad212\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:06Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:06 crc kubenswrapper[4967]: I1008 00:07:06.448425 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9tdbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41264a12-9486-4fee-a08c-607b590e5a70\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://806e85df796ae300d279f73a0cc4433a5bb1f7620ee8ce33c8dc4674c4ce0596\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5blv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9tdbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:06Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:06 crc kubenswrapper[4967]: I1008 00:07:06.466990 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f1af2ba-8907-4d14-afa1-ffde207c80f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bs8w9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:06Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:06 crc kubenswrapper[4967]: I1008 00:07:06.475914 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-v5n7t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abe7ae2d-7e63-4249-96e1-4000e9628ccc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpsjd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:06Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-v5n7t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:06Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:06 crc kubenswrapper[4967]: I1008 00:07:06.489121 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:06Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:06 crc kubenswrapper[4967]: I1008 00:07:06.500424 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-npz2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c71501b1-aa9e-420e-b1bd-7e15365e7adc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dfe6c420d804b2d7daef4b5cfe3f87b77b369287c4d65553fa93f5f24f589d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gblk2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-npz2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:06Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:06 crc kubenswrapper[4967]: I1008 00:07:06.513059 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58c4704-a88a-43b4-b944-4de626f13f7e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22f5dabb39da0a7bd01db4dae1437b9a231a1fcc0dfe749a4a5a637a72cf33c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01a2b3dc785520c79eae998482e5a1347725c76584357073238fe60a619edb77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6aa6f5e81feceab3d1c7587152a70546b365e629a69267e4d3084324a1de764a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63c2b7c4b2791bd1f16888c58b878dbff1f87884e68c9949de914d0d3e1d39fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:06Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:06 crc kubenswrapper[4967]: I1008 00:07:06.524201 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:06Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:06 crc kubenswrapper[4967]: I1008 00:07:06.539173 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9mrr7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9299bc4f-f8d3-4baf-9885-39573982fd64\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb49c8476b482416588c4d403016940730064a619c0641eb029c947a4bab1215\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb49c8476b482416588c4d403016940730064a619c0641eb029c947a4bab1215\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9mrr7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:06Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:06 crc kubenswrapper[4967]: I1008 00:07:06.552274 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fa4b747-ab6e-4a11-9903-ef2dcb9668c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbadcdade274b6d9be940f22bc949e3c55cadf95dcad43c8cb7a1085c0e6708c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d0a0b2ef4a98ee828688081a74f4b7ecc31c8fae415d7257a560ecd43960ce3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96b53b1b283025b3d078a3f505d469d1872ad4bb8ab8e612c36c70b0b00a920\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1400e110997c64fa33d6b8672c7e91cab5f782952ddf8e1dea2a882689026e7f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1400e110997c64fa33d6b8672c7e91cab5f782952ddf8e1dea2a882689026e7f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 00:06:56.760088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 00:06:56.761210 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-369805205/tls.crt::/tmp/serving-cert-369805205/tls.key\\\\\\\"\\\\nI1008 00:07:02.321448 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 00:07:02.325932 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 00:07:02.325980 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 00:07:02.326025 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 00:07:02.326042 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 00:07:02.336203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 00:07:02.336240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 00:07:02.336245 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 00:07:02.336250 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 00:07:02.336253 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 00:07:02.336256 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 00:07:02.336258 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 00:07:02.336430 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 00:07:02.340774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:46Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca28d8db19036e47a7ef3dadb70941f6c6cf5735e7703977d036185c46fce40\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5b222ad84c31d8746dfe03124c7287b401deb3ca2c62ce2b4fbd40e638cee0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5b222ad84c31d8746dfe03124c7287b401deb3ca2c62ce2b4fbd40e638cee0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:06Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:06 crc kubenswrapper[4967]: I1008 00:07:06.565215 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:06Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:06 crc kubenswrapper[4967]: I1008 00:07:06.581022 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbfcdc882e2be1cb9cf666c9e7775671078f7dbbfe8421dd15db737b89920a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:06Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:06 crc kubenswrapper[4967]: I1008 00:07:06.591591 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9461162a-c27d-465e-9ca7-2d3936877cd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9778d78ddc7325abd4d77bc5e71b0b114f50be5dda4a438c7cb6b05640c8a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z9tp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28760af04ef321d0f91b2bce3ad074d44b63dabf676cc37f3be6aa8726341a2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z9tp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xvdtp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:06Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:06 crc kubenswrapper[4967]: I1008 00:07:06.631815 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-v5n7t" Oct 08 00:07:06 crc kubenswrapper[4967]: W1008 00:07:06.650526 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podabe7ae2d_7e63_4249_96e1_4000e9628ccc.slice/crio-4b76a14f851ac02e3533cb26f9518116bd01758eb11f57a4bcf043e13662deb8 WatchSource:0}: Error finding container 4b76a14f851ac02e3533cb26f9518116bd01758eb11f57a4bcf043e13662deb8: Status 404 returned error can't find the container with id 4b76a14f851ac02e3533cb26f9518116bd01758eb11f57a4bcf043e13662deb8 Oct 08 00:07:06 crc kubenswrapper[4967]: I1008 00:07:06.892749 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 00:07:06 crc kubenswrapper[4967]: E1008 00:07:06.893690 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 00:07:10.893026386 +0000 UTC m=+29.543561035 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:07:06 crc kubenswrapper[4967]: I1008 00:07:06.893794 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 00:07:06 crc kubenswrapper[4967]: I1008 00:07:06.893896 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 00:07:06 crc kubenswrapper[4967]: I1008 00:07:06.893983 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 00:07:06 crc kubenswrapper[4967]: I1008 00:07:06.894028 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 00:07:06 crc kubenswrapper[4967]: E1008 00:07:06.893903 4967 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 00:07:06 crc kubenswrapper[4967]: E1008 00:07:06.894144 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 00:07:06 crc kubenswrapper[4967]: E1008 00:07:06.894172 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 00:07:06 crc kubenswrapper[4967]: E1008 00:07:06.894221 4967 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 00:07:06 crc kubenswrapper[4967]: E1008 00:07:06.894233 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 00:07:10.894187516 +0000 UTC m=+29.544722175 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 00:07:06 crc kubenswrapper[4967]: E1008 00:07:06.894273 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-08 00:07:10.894253569 +0000 UTC m=+29.544788208 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 00:07:06 crc kubenswrapper[4967]: E1008 00:07:06.894035 4967 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 00:07:06 crc kubenswrapper[4967]: E1008 00:07:06.894305 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 00:07:10.89429916 +0000 UTC m=+29.544833809 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 00:07:06 crc kubenswrapper[4967]: E1008 00:07:06.895168 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 00:07:06 crc kubenswrapper[4967]: E1008 00:07:06.895220 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 00:07:06 crc kubenswrapper[4967]: E1008 00:07:06.895303 4967 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 00:07:06 crc kubenswrapper[4967]: E1008 00:07:06.895399 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-08 00:07:10.895383828 +0000 UTC m=+29.545918647 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 00:07:07 crc kubenswrapper[4967]: I1008 00:07:07.117389 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 00:07:07 crc kubenswrapper[4967]: E1008 00:07:07.117532 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 00:07:07 crc kubenswrapper[4967]: I1008 00:07:07.117889 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 00:07:07 crc kubenswrapper[4967]: E1008 00:07:07.118039 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 00:07:07 crc kubenswrapper[4967]: I1008 00:07:07.370157 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-v5n7t" event={"ID":"abe7ae2d-7e63-4249-96e1-4000e9628ccc","Type":"ContainerStarted","Data":"52575e853ff6727dde7c89e4add4362be126127db27191c2a7bf2871efa719d6"} Oct 08 00:07:07 crc kubenswrapper[4967]: I1008 00:07:07.370236 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-v5n7t" event={"ID":"abe7ae2d-7e63-4249-96e1-4000e9628ccc","Type":"ContainerStarted","Data":"4b76a14f851ac02e3533cb26f9518116bd01758eb11f57a4bcf043e13662deb8"} Oct 08 00:07:07 crc kubenswrapper[4967]: I1008 00:07:07.373087 4967 generic.go:334] "Generic (PLEG): container finished" podID="9299bc4f-f8d3-4baf-9885-39573982fd64" containerID="3b7b63962dcbf477f1fa4d71826a4aa3358920a2c1b2f5e48716411b5bdca049" exitCode=0 Oct 08 00:07:07 crc kubenswrapper[4967]: I1008 00:07:07.373184 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9mrr7" event={"ID":"9299bc4f-f8d3-4baf-9885-39573982fd64","Type":"ContainerDied","Data":"3b7b63962dcbf477f1fa4d71826a4aa3358920a2c1b2f5e48716411b5bdca049"} Oct 08 00:07:07 crc kubenswrapper[4967]: I1008 00:07:07.389783 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:07Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:07 crc kubenswrapper[4967]: I1008 00:07:07.404570 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-npz2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c71501b1-aa9e-420e-b1bd-7e15365e7adc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dfe6c420d804b2d7daef4b5cfe3f87b77b369287c4d65553fa93f5f24f589d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gblk2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-npz2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:07Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:07 crc kubenswrapper[4967]: I1008 00:07:07.419993 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58c4704-a88a-43b4-b944-4de626f13f7e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22f5dabb39da0a7bd01db4dae1437b9a231a1fcc0dfe749a4a5a637a72cf33c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01a2b3dc785520c79eae998482e5a1347725c76584357073238fe60a619edb77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6aa6f5e81feceab3d1c7587152a70546b365e629a69267e4d3084324a1de764a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63c2b7c4b2791bd1f16888c58b878dbff1f87884e68c9949de914d0d3e1d39fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:07Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:07 crc kubenswrapper[4967]: I1008 00:07:07.435039 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:07Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:07 crc kubenswrapper[4967]: I1008 00:07:07.449790 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9mrr7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9299bc4f-f8d3-4baf-9885-39573982fd64\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb49c8476b482416588c4d403016940730064a619c0641eb029c947a4bab1215\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb49c8476b482416588c4d403016940730064a619c0641eb029c947a4bab1215\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9mrr7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:07Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:07 crc kubenswrapper[4967]: I1008 00:07:07.465857 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fa4b747-ab6e-4a11-9903-ef2dcb9668c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbadcdade274b6d9be940f22bc949e3c55cadf95dcad43c8cb7a1085c0e6708c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d0a0b2ef4a98ee828688081a74f4b7ecc31c8fae415d7257a560ecd43960ce3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96b53b1b283025b3d078a3f505d469d1872ad4bb8ab8e612c36c70b0b00a920\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1400e110997c64fa33d6b8672c7e91cab5f782952ddf8e1dea2a882689026e7f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1400e110997c64fa33d6b8672c7e91cab5f782952ddf8e1dea2a882689026e7f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 00:06:56.760088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 00:06:56.761210 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-369805205/tls.crt::/tmp/serving-cert-369805205/tls.key\\\\\\\"\\\\nI1008 00:07:02.321448 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 00:07:02.325932 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 00:07:02.325980 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 00:07:02.326025 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 00:07:02.326042 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 00:07:02.336203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 00:07:02.336240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 00:07:02.336245 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 00:07:02.336250 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 00:07:02.336253 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 00:07:02.336256 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 00:07:02.336258 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 00:07:02.336430 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 00:07:02.340774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:46Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca28d8db19036e47a7ef3dadb70941f6c6cf5735e7703977d036185c46fce40\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5b222ad84c31d8746dfe03124c7287b401deb3ca2c62ce2b4fbd40e638cee0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5b222ad84c31d8746dfe03124c7287b401deb3ca2c62ce2b4fbd40e638cee0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:07Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:07 crc kubenswrapper[4967]: I1008 00:07:07.479024 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:07Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:07 crc kubenswrapper[4967]: I1008 00:07:07.488634 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbfcdc882e2be1cb9cf666c9e7775671078f7dbbfe8421dd15db737b89920a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:07Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:07 crc kubenswrapper[4967]: I1008 00:07:07.499029 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9461162a-c27d-465e-9ca7-2d3936877cd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9778d78ddc7325abd4d77bc5e71b0b114f50be5dda4a438c7cb6b05640c8a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z9tp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28760af04ef321d0f91b2bce3ad074d44b63dabf676cc37f3be6aa8726341a2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z9tp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xvdtp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:07Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:07 crc kubenswrapper[4967]: I1008 00:07:07.513244 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94e63bc5641ce80d507f76a19b203f78c42368a7a4105f3fd09e19af21a1b735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:07Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:07 crc kubenswrapper[4967]: I1008 00:07:07.528414 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e98666fd3152bab57d3723e7da72aae2703b79f7ca73eb3c3e5f0e062a82592\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cf1cbe5548146fe359aafa351864125ec6b8bdd7b796aa4d311f6f3a16ad212\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:07Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:07 crc kubenswrapper[4967]: I1008 00:07:07.538435 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9tdbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41264a12-9486-4fee-a08c-607b590e5a70\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://806e85df796ae300d279f73a0cc4433a5bb1f7620ee8ce33c8dc4674c4ce0596\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5blv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9tdbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:07Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:07 crc kubenswrapper[4967]: I1008 00:07:07.563622 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f1af2ba-8907-4d14-afa1-ffde207c80f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bs8w9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:07Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:07 crc kubenswrapper[4967]: I1008 00:07:07.573927 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-v5n7t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abe7ae2d-7e63-4249-96e1-4000e9628ccc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52575e853ff6727dde7c89e4add4362be126127db27191c2a7bf2871efa719d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpsjd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:06Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-v5n7t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:07Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:07 crc kubenswrapper[4967]: I1008 00:07:07.586712 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:07Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:07 crc kubenswrapper[4967]: I1008 00:07:07.599380 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbfcdc882e2be1cb9cf666c9e7775671078f7dbbfe8421dd15db737b89920a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:07Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:07 crc kubenswrapper[4967]: I1008 00:07:07.610192 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9461162a-c27d-465e-9ca7-2d3936877cd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9778d78ddc7325abd4d77bc5e71b0b114f50be5dda4a438c7cb6b05640c8a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z9tp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28760af04ef321d0f91b2bce3ad074d44b63dabf676cc37f3be6aa8726341a2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z9tp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xvdtp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:07Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:07 crc kubenswrapper[4967]: I1008 00:07:07.623152 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fa4b747-ab6e-4a11-9903-ef2dcb9668c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbadcdade274b6d9be940f22bc949e3c55cadf95dcad43c8cb7a1085c0e6708c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d0a0b2ef4a98ee828688081a74f4b7ecc31c8fae415d7257a560ecd43960ce3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96b53b1b283025b3d078a3f505d469d1872ad4bb8ab8e612c36c70b0b00a920\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1400e110997c64fa33d6b8672c7e91cab5f782952ddf8e1dea2a882689026e7f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1400e110997c64fa33d6b8672c7e91cab5f782952ddf8e1dea2a882689026e7f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 00:06:56.760088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 00:06:56.761210 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-369805205/tls.crt::/tmp/serving-cert-369805205/tls.key\\\\\\\"\\\\nI1008 00:07:02.321448 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 00:07:02.325932 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 00:07:02.325980 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 00:07:02.326025 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 00:07:02.326042 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 00:07:02.336203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 00:07:02.336240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 00:07:02.336245 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 00:07:02.336250 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 00:07:02.336253 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 00:07:02.336256 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 00:07:02.336258 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 00:07:02.336430 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 00:07:02.340774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:46Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca28d8db19036e47a7ef3dadb70941f6c6cf5735e7703977d036185c46fce40\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5b222ad84c31d8746dfe03124c7287b401deb3ca2c62ce2b4fbd40e638cee0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5b222ad84c31d8746dfe03124c7287b401deb3ca2c62ce2b4fbd40e638cee0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:07Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:07 crc kubenswrapper[4967]: I1008 00:07:07.635389 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9tdbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41264a12-9486-4fee-a08c-607b590e5a70\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://806e85df796ae300d279f73a0cc4433a5bb1f7620ee8ce33c8dc4674c4ce0596\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5blv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9tdbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:07Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:07 crc kubenswrapper[4967]: I1008 00:07:07.655119 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f1af2ba-8907-4d14-afa1-ffde207c80f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bs8w9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:07Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:07 crc kubenswrapper[4967]: I1008 00:07:07.664531 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-v5n7t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abe7ae2d-7e63-4249-96e1-4000e9628ccc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52575e853ff6727dde7c89e4add4362be126127db27191c2a7bf2871efa719d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpsjd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:06Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-v5n7t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:07Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:07 crc kubenswrapper[4967]: I1008 00:07:07.675877 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94e63bc5641ce80d507f76a19b203f78c42368a7a4105f3fd09e19af21a1b735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:07Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:07 crc kubenswrapper[4967]: I1008 00:07:07.686688 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e98666fd3152bab57d3723e7da72aae2703b79f7ca73eb3c3e5f0e062a82592\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cf1cbe5548146fe359aafa351864125ec6b8bdd7b796aa4d311f6f3a16ad212\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:07Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:07 crc kubenswrapper[4967]: I1008 00:07:07.698557 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-npz2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c71501b1-aa9e-420e-b1bd-7e15365e7adc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dfe6c420d804b2d7daef4b5cfe3f87b77b369287c4d65553fa93f5f24f589d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gblk2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-npz2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:07Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:07 crc kubenswrapper[4967]: I1008 00:07:07.711935 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:07Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:07 crc kubenswrapper[4967]: I1008 00:07:07.725295 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:07Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:07 crc kubenswrapper[4967]: I1008 00:07:07.743181 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9mrr7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9299bc4f-f8d3-4baf-9885-39573982fd64\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb49c8476b482416588c4d403016940730064a619c0641eb029c947a4bab1215\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb49c8476b482416588c4d403016940730064a619c0641eb029c947a4bab1215\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b7b63962dcbf477f1fa4d71826a4aa3358920a2c1b2f5e48716411b5bdca049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b7b63962dcbf477f1fa4d71826a4aa3358920a2c1b2f5e48716411b5bdca049\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9mrr7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:07Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:07 crc kubenswrapper[4967]: I1008 00:07:07.756922 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58c4704-a88a-43b4-b944-4de626f13f7e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22f5dabb39da0a7bd01db4dae1437b9a231a1fcc0dfe749a4a5a637a72cf33c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01a2b3dc785520c79eae998482e5a1347725c76584357073238fe60a619edb77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6aa6f5e81feceab3d1c7587152a70546b365e629a69267e4d3084324a1de764a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63c2b7c4b2791bd1f16888c58b878dbff1f87884e68c9949de914d0d3e1d39fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:07Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:08 crc kubenswrapper[4967]: I1008 00:07:08.117516 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 00:07:08 crc kubenswrapper[4967]: E1008 00:07:08.117833 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 00:07:08 crc kubenswrapper[4967]: I1008 00:07:08.379030 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" event={"ID":"4f1af2ba-8907-4d14-afa1-ffde207c80f4","Type":"ContainerStarted","Data":"27fb2b1225e386dcd7fecaaa0aa54ba1fdfafabfde6c3824ae07315313067542"} Oct 08 00:07:08 crc kubenswrapper[4967]: I1008 00:07:08.381072 4967 generic.go:334] "Generic (PLEG): container finished" podID="9299bc4f-f8d3-4baf-9885-39573982fd64" containerID="d1e770dc737173fd4f3220ed6c52b422ccbd829722386a37029e1c3e10d80e1a" exitCode=0 Oct 08 00:07:08 crc kubenswrapper[4967]: I1008 00:07:08.381121 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9mrr7" event={"ID":"9299bc4f-f8d3-4baf-9885-39573982fd64","Type":"ContainerDied","Data":"d1e770dc737173fd4f3220ed6c52b422ccbd829722386a37029e1c3e10d80e1a"} Oct 08 00:07:08 crc kubenswrapper[4967]: I1008 00:07:08.400796 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:08Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:08 crc kubenswrapper[4967]: I1008 00:07:08.415973 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9mrr7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9299bc4f-f8d3-4baf-9885-39573982fd64\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb49c8476b482416588c4d403016940730064a619c0641eb029c947a4bab1215\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb49c8476b482416588c4d403016940730064a619c0641eb029c947a4bab1215\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b7b63962dcbf477f1fa4d71826a4aa3358920a2c1b2f5e48716411b5bdca049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b7b63962dcbf477f1fa4d71826a4aa3358920a2c1b2f5e48716411b5bdca049\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1e770dc737173fd4f3220ed6c52b422ccbd829722386a37029e1c3e10d80e1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1e770dc737173fd4f3220ed6c52b422ccbd829722386a37029e1c3e10d80e1a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9mrr7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:08Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:08 crc kubenswrapper[4967]: I1008 00:07:08.428918 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58c4704-a88a-43b4-b944-4de626f13f7e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22f5dabb39da0a7bd01db4dae1437b9a231a1fcc0dfe749a4a5a637a72cf33c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01a2b3dc785520c79eae998482e5a1347725c76584357073238fe60a619edb77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6aa6f5e81feceab3d1c7587152a70546b365e629a69267e4d3084324a1de764a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63c2b7c4b2791bd1f16888c58b878dbff1f87884e68c9949de914d0d3e1d39fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:08Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:08 crc kubenswrapper[4967]: I1008 00:07:08.445545 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:08Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:08 crc kubenswrapper[4967]: I1008 00:07:08.459795 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbfcdc882e2be1cb9cf666c9e7775671078f7dbbfe8421dd15db737b89920a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:08Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:08 crc kubenswrapper[4967]: I1008 00:07:08.473541 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9461162a-c27d-465e-9ca7-2d3936877cd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9778d78ddc7325abd4d77bc5e71b0b114f50be5dda4a438c7cb6b05640c8a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z9tp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28760af04ef321d0f91b2bce3ad074d44b63dabf676cc37f3be6aa8726341a2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z9tp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xvdtp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:08Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:08 crc kubenswrapper[4967]: I1008 00:07:08.487632 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fa4b747-ab6e-4a11-9903-ef2dcb9668c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbadcdade274b6d9be940f22bc949e3c55cadf95dcad43c8cb7a1085c0e6708c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d0a0b2ef4a98ee828688081a74f4b7ecc31c8fae415d7257a560ecd43960ce3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96b53b1b283025b3d078a3f505d469d1872ad4bb8ab8e612c36c70b0b00a920\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1400e110997c64fa33d6b8672c7e91cab5f782952ddf8e1dea2a882689026e7f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1400e110997c64fa33d6b8672c7e91cab5f782952ddf8e1dea2a882689026e7f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 00:06:56.760088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 00:06:56.761210 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-369805205/tls.crt::/tmp/serving-cert-369805205/tls.key\\\\\\\"\\\\nI1008 00:07:02.321448 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 00:07:02.325932 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 00:07:02.325980 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 00:07:02.326025 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 00:07:02.326042 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 00:07:02.336203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 00:07:02.336240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 00:07:02.336245 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 00:07:02.336250 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 00:07:02.336253 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 00:07:02.336256 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 00:07:02.336258 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 00:07:02.336430 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 00:07:02.340774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:46Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca28d8db19036e47a7ef3dadb70941f6c6cf5735e7703977d036185c46fce40\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5b222ad84c31d8746dfe03124c7287b401deb3ca2c62ce2b4fbd40e638cee0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5b222ad84c31d8746dfe03124c7287b401deb3ca2c62ce2b4fbd40e638cee0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:08Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:08 crc kubenswrapper[4967]: I1008 00:07:08.499914 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9tdbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41264a12-9486-4fee-a08c-607b590e5a70\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://806e85df796ae300d279f73a0cc4433a5bb1f7620ee8ce33c8dc4674c4ce0596\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5blv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9tdbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:08Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:08 crc kubenswrapper[4967]: I1008 00:07:08.519785 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f1af2ba-8907-4d14-afa1-ffde207c80f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bs8w9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:08Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:08 crc kubenswrapper[4967]: I1008 00:07:08.529433 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-v5n7t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abe7ae2d-7e63-4249-96e1-4000e9628ccc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52575e853ff6727dde7c89e4add4362be126127db27191c2a7bf2871efa719d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpsjd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:06Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-v5n7t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:08Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:08 crc kubenswrapper[4967]: I1008 00:07:08.540576 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94e63bc5641ce80d507f76a19b203f78c42368a7a4105f3fd09e19af21a1b735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:08Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:08 crc kubenswrapper[4967]: I1008 00:07:08.551875 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e98666fd3152bab57d3723e7da72aae2703b79f7ca73eb3c3e5f0e062a82592\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cf1cbe5548146fe359aafa351864125ec6b8bdd7b796aa4d311f6f3a16ad212\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:08Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:08 crc kubenswrapper[4967]: I1008 00:07:08.565208 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-npz2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c71501b1-aa9e-420e-b1bd-7e15365e7adc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dfe6c420d804b2d7daef4b5cfe3f87b77b369287c4d65553fa93f5f24f589d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gblk2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-npz2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:08Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:08 crc kubenswrapper[4967]: I1008 00:07:08.579857 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:08Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:08 crc kubenswrapper[4967]: I1008 00:07:08.698923 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 00:07:08 crc kubenswrapper[4967]: I1008 00:07:08.702619 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:08 crc kubenswrapper[4967]: I1008 00:07:08.702678 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:08 crc kubenswrapper[4967]: I1008 00:07:08.702690 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:08 crc kubenswrapper[4967]: I1008 00:07:08.702790 4967 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 08 00:07:08 crc kubenswrapper[4967]: I1008 00:07:08.710138 4967 kubelet_node_status.go:115] "Node was previously registered" node="crc" Oct 08 00:07:08 crc kubenswrapper[4967]: I1008 00:07:08.710369 4967 kubelet_node_status.go:79] "Successfully registered node" node="crc" Oct 08 00:07:08 crc kubenswrapper[4967]: I1008 00:07:08.713290 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:08 crc kubenswrapper[4967]: I1008 00:07:08.713485 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:08 crc kubenswrapper[4967]: I1008 00:07:08.713511 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:08 crc kubenswrapper[4967]: I1008 00:07:08.713548 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:08 crc kubenswrapper[4967]: I1008 00:07:08.713573 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:08Z","lastTransitionTime":"2025-10-08T00:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:08 crc kubenswrapper[4967]: E1008 00:07:08.731738 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5cddad4c-22d1-42d7-a373-87b47c1cc97b\\\",\\\"systemUUID\\\":\\\"b243ca1c-20cd-4e6f-96b2-52fce74dd2d4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:08Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:08 crc kubenswrapper[4967]: I1008 00:07:08.735207 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:08 crc kubenswrapper[4967]: I1008 00:07:08.735238 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:08 crc kubenswrapper[4967]: I1008 00:07:08.735247 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:08 crc kubenswrapper[4967]: I1008 00:07:08.735259 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:08 crc kubenswrapper[4967]: I1008 00:07:08.735267 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:08Z","lastTransitionTime":"2025-10-08T00:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:08 crc kubenswrapper[4967]: E1008 00:07:08.748258 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5cddad4c-22d1-42d7-a373-87b47c1cc97b\\\",\\\"systemUUID\\\":\\\"b243ca1c-20cd-4e6f-96b2-52fce74dd2d4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:08Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:08 crc kubenswrapper[4967]: I1008 00:07:08.752323 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:08 crc kubenswrapper[4967]: I1008 00:07:08.752360 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:08 crc kubenswrapper[4967]: I1008 00:07:08.752372 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:08 crc kubenswrapper[4967]: I1008 00:07:08.752387 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:08 crc kubenswrapper[4967]: I1008 00:07:08.752401 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:08Z","lastTransitionTime":"2025-10-08T00:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:08 crc kubenswrapper[4967]: E1008 00:07:08.767046 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5cddad4c-22d1-42d7-a373-87b47c1cc97b\\\",\\\"systemUUID\\\":\\\"b243ca1c-20cd-4e6f-96b2-52fce74dd2d4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:08Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:08 crc kubenswrapper[4967]: I1008 00:07:08.771195 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:08 crc kubenswrapper[4967]: I1008 00:07:08.771226 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:08 crc kubenswrapper[4967]: I1008 00:07:08.771237 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:08 crc kubenswrapper[4967]: I1008 00:07:08.771325 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:08 crc kubenswrapper[4967]: I1008 00:07:08.771399 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:08Z","lastTransitionTime":"2025-10-08T00:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:08 crc kubenswrapper[4967]: E1008 00:07:08.784571 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5cddad4c-22d1-42d7-a373-87b47c1cc97b\\\",\\\"systemUUID\\\":\\\"b243ca1c-20cd-4e6f-96b2-52fce74dd2d4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:08Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:08 crc kubenswrapper[4967]: I1008 00:07:08.787783 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:08 crc kubenswrapper[4967]: I1008 00:07:08.787805 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:08 crc kubenswrapper[4967]: I1008 00:07:08.787816 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:08 crc kubenswrapper[4967]: I1008 00:07:08.788170 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:08 crc kubenswrapper[4967]: I1008 00:07:08.788184 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:08Z","lastTransitionTime":"2025-10-08T00:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:08 crc kubenswrapper[4967]: E1008 00:07:08.844210 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5cddad4c-22d1-42d7-a373-87b47c1cc97b\\\",\\\"systemUUID\\\":\\\"b243ca1c-20cd-4e6f-96b2-52fce74dd2d4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:08Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:08 crc kubenswrapper[4967]: E1008 00:07:08.844364 4967 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 08 00:07:08 crc kubenswrapper[4967]: I1008 00:07:08.845791 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:08 crc kubenswrapper[4967]: I1008 00:07:08.845815 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:08 crc kubenswrapper[4967]: I1008 00:07:08.845825 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:08 crc kubenswrapper[4967]: I1008 00:07:08.845842 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:08 crc kubenswrapper[4967]: I1008 00:07:08.845853 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:08Z","lastTransitionTime":"2025-10-08T00:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:08 crc kubenswrapper[4967]: I1008 00:07:08.947721 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:08 crc kubenswrapper[4967]: I1008 00:07:08.947768 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:08 crc kubenswrapper[4967]: I1008 00:07:08.947781 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:08 crc kubenswrapper[4967]: I1008 00:07:08.947803 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:08 crc kubenswrapper[4967]: I1008 00:07:08.947815 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:08Z","lastTransitionTime":"2025-10-08T00:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:09 crc kubenswrapper[4967]: I1008 00:07:09.050162 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:09 crc kubenswrapper[4967]: I1008 00:07:09.050193 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:09 crc kubenswrapper[4967]: I1008 00:07:09.050203 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:09 crc kubenswrapper[4967]: I1008 00:07:09.050217 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:09 crc kubenswrapper[4967]: I1008 00:07:09.050226 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:09Z","lastTransitionTime":"2025-10-08T00:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:09 crc kubenswrapper[4967]: I1008 00:07:09.117787 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 00:07:09 crc kubenswrapper[4967]: E1008 00:07:09.117916 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 00:07:09 crc kubenswrapper[4967]: I1008 00:07:09.118244 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 00:07:09 crc kubenswrapper[4967]: E1008 00:07:09.118299 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 00:07:09 crc kubenswrapper[4967]: I1008 00:07:09.152258 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:09 crc kubenswrapper[4967]: I1008 00:07:09.152289 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:09 crc kubenswrapper[4967]: I1008 00:07:09.152300 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:09 crc kubenswrapper[4967]: I1008 00:07:09.152312 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:09 crc kubenswrapper[4967]: I1008 00:07:09.152320 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:09Z","lastTransitionTime":"2025-10-08T00:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:09 crc kubenswrapper[4967]: I1008 00:07:09.254924 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:09 crc kubenswrapper[4967]: I1008 00:07:09.254955 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:09 crc kubenswrapper[4967]: I1008 00:07:09.254967 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:09 crc kubenswrapper[4967]: I1008 00:07:09.254981 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:09 crc kubenswrapper[4967]: I1008 00:07:09.254991 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:09Z","lastTransitionTime":"2025-10-08T00:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:09 crc kubenswrapper[4967]: I1008 00:07:09.357431 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:09 crc kubenswrapper[4967]: I1008 00:07:09.357462 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:09 crc kubenswrapper[4967]: I1008 00:07:09.357492 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:09 crc kubenswrapper[4967]: I1008 00:07:09.357508 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:09 crc kubenswrapper[4967]: I1008 00:07:09.357520 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:09Z","lastTransitionTime":"2025-10-08T00:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:09 crc kubenswrapper[4967]: I1008 00:07:09.387153 4967 generic.go:334] "Generic (PLEG): container finished" podID="9299bc4f-f8d3-4baf-9885-39573982fd64" containerID="e357d970878c996f219a79f1876b363176c9847b9575f33090f62bc299529a95" exitCode=0 Oct 08 00:07:09 crc kubenswrapper[4967]: I1008 00:07:09.387203 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9mrr7" event={"ID":"9299bc4f-f8d3-4baf-9885-39573982fd64","Type":"ContainerDied","Data":"e357d970878c996f219a79f1876b363176c9847b9575f33090f62bc299529a95"} Oct 08 00:07:09 crc kubenswrapper[4967]: I1008 00:07:09.410205 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f1af2ba-8907-4d14-afa1-ffde207c80f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bs8w9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:09Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:09 crc kubenswrapper[4967]: I1008 00:07:09.422828 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-v5n7t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abe7ae2d-7e63-4249-96e1-4000e9628ccc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52575e853ff6727dde7c89e4add4362be126127db27191c2a7bf2871efa719d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpsjd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:06Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-v5n7t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:09Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:09 crc kubenswrapper[4967]: I1008 00:07:09.445127 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94e63bc5641ce80d507f76a19b203f78c42368a7a4105f3fd09e19af21a1b735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:09Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:09 crc kubenswrapper[4967]: I1008 00:07:09.458757 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e98666fd3152bab57d3723e7da72aae2703b79f7ca73eb3c3e5f0e062a82592\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cf1cbe5548146fe359aafa351864125ec6b8bdd7b796aa4d311f6f3a16ad212\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:09Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:09 crc kubenswrapper[4967]: I1008 00:07:09.459661 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:09 crc kubenswrapper[4967]: I1008 00:07:09.459743 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:09 crc kubenswrapper[4967]: I1008 00:07:09.459767 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:09 crc kubenswrapper[4967]: I1008 00:07:09.459795 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:09 crc kubenswrapper[4967]: I1008 00:07:09.459817 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:09Z","lastTransitionTime":"2025-10-08T00:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:09 crc kubenswrapper[4967]: I1008 00:07:09.469810 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9tdbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41264a12-9486-4fee-a08c-607b590e5a70\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://806e85df796ae300d279f73a0cc4433a5bb1f7620ee8ce33c8dc4674c4ce0596\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5blv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9tdbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:09Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:09 crc kubenswrapper[4967]: I1008 00:07:09.484112 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:09Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:09 crc kubenswrapper[4967]: I1008 00:07:09.501086 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-npz2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c71501b1-aa9e-420e-b1bd-7e15365e7adc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dfe6c420d804b2d7daef4b5cfe3f87b77b369287c4d65553fa93f5f24f589d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gblk2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-npz2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:09Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:09 crc kubenswrapper[4967]: I1008 00:07:09.518594 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9mrr7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9299bc4f-f8d3-4baf-9885-39573982fd64\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb49c8476b482416588c4d403016940730064a619c0641eb029c947a4bab1215\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb49c8476b482416588c4d403016940730064a619c0641eb029c947a4bab1215\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b7b63962dcbf477f1fa4d71826a4aa3358920a2c1b2f5e48716411b5bdca049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b7b63962dcbf477f1fa4d71826a4aa3358920a2c1b2f5e48716411b5bdca049\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1e770dc737173fd4f3220ed6c52b422ccbd829722386a37029e1c3e10d80e1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1e770dc737173fd4f3220ed6c52b422ccbd829722386a37029e1c3e10d80e1a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e357d970878c996f219a79f1876b363176c9847b9575f33090f62bc299529a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e357d970878c996f219a79f1876b363176c9847b9575f33090f62bc299529a95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9mrr7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:09Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:09 crc kubenswrapper[4967]: I1008 00:07:09.536350 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58c4704-a88a-43b4-b944-4de626f13f7e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22f5dabb39da0a7bd01db4dae1437b9a231a1fcc0dfe749a4a5a637a72cf33c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01a2b3dc785520c79eae998482e5a1347725c76584357073238fe60a619edb77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6aa6f5e81feceab3d1c7587152a70546b365e629a69267e4d3084324a1de764a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63c2b7c4b2791bd1f16888c58b878dbff1f87884e68c9949de914d0d3e1d39fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:09Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:09 crc kubenswrapper[4967]: I1008 00:07:09.549726 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:09Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:09 crc kubenswrapper[4967]: I1008 00:07:09.560260 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbfcdc882e2be1cb9cf666c9e7775671078f7dbbfe8421dd15db737b89920a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:09Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:09 crc kubenswrapper[4967]: I1008 00:07:09.566944 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:09 crc kubenswrapper[4967]: I1008 00:07:09.566998 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:09 crc kubenswrapper[4967]: I1008 00:07:09.567014 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:09 crc kubenswrapper[4967]: I1008 00:07:09.567046 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:09 crc kubenswrapper[4967]: I1008 00:07:09.567064 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:09Z","lastTransitionTime":"2025-10-08T00:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:09 crc kubenswrapper[4967]: I1008 00:07:09.572879 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9461162a-c27d-465e-9ca7-2d3936877cd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9778d78ddc7325abd4d77bc5e71b0b114f50be5dda4a438c7cb6b05640c8a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z9tp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28760af04ef321d0f91b2bce3ad074d44b63dabf676cc37f3be6aa8726341a2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z9tp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xvdtp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:09Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:09 crc kubenswrapper[4967]: I1008 00:07:09.586361 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fa4b747-ab6e-4a11-9903-ef2dcb9668c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbadcdade274b6d9be940f22bc949e3c55cadf95dcad43c8cb7a1085c0e6708c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d0a0b2ef4a98ee828688081a74f4b7ecc31c8fae415d7257a560ecd43960ce3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96b53b1b283025b3d078a3f505d469d1872ad4bb8ab8e612c36c70b0b00a920\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1400e110997c64fa33d6b8672c7e91cab5f782952ddf8e1dea2a882689026e7f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1400e110997c64fa33d6b8672c7e91cab5f782952ddf8e1dea2a882689026e7f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 00:06:56.760088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 00:06:56.761210 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-369805205/tls.crt::/tmp/serving-cert-369805205/tls.key\\\\\\\"\\\\nI1008 00:07:02.321448 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 00:07:02.325932 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 00:07:02.325980 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 00:07:02.326025 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 00:07:02.326042 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 00:07:02.336203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 00:07:02.336240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 00:07:02.336245 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 00:07:02.336250 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 00:07:02.336253 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 00:07:02.336256 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 00:07:02.336258 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 00:07:02.336430 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 00:07:02.340774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:46Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca28d8db19036e47a7ef3dadb70941f6c6cf5735e7703977d036185c46fce40\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5b222ad84c31d8746dfe03124c7287b401deb3ca2c62ce2b4fbd40e638cee0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5b222ad84c31d8746dfe03124c7287b401deb3ca2c62ce2b4fbd40e638cee0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:09Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:09 crc kubenswrapper[4967]: I1008 00:07:09.600887 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:09Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:09 crc kubenswrapper[4967]: I1008 00:07:09.669407 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:09 crc kubenswrapper[4967]: I1008 00:07:09.669474 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:09 crc kubenswrapper[4967]: I1008 00:07:09.669485 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:09 crc kubenswrapper[4967]: I1008 00:07:09.669504 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:09 crc kubenswrapper[4967]: I1008 00:07:09.669516 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:09Z","lastTransitionTime":"2025-10-08T00:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:09 crc kubenswrapper[4967]: I1008 00:07:09.773000 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:09 crc kubenswrapper[4967]: I1008 00:07:09.773055 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:09 crc kubenswrapper[4967]: I1008 00:07:09.773072 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:09 crc kubenswrapper[4967]: I1008 00:07:09.773096 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:09 crc kubenswrapper[4967]: I1008 00:07:09.773113 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:09Z","lastTransitionTime":"2025-10-08T00:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:09 crc kubenswrapper[4967]: I1008 00:07:09.876608 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:09 crc kubenswrapper[4967]: I1008 00:07:09.876715 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:09 crc kubenswrapper[4967]: I1008 00:07:09.876737 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:09 crc kubenswrapper[4967]: I1008 00:07:09.876761 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:09 crc kubenswrapper[4967]: I1008 00:07:09.876812 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:09Z","lastTransitionTime":"2025-10-08T00:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:09 crc kubenswrapper[4967]: I1008 00:07:09.979768 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:09 crc kubenswrapper[4967]: I1008 00:07:09.979814 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:09 crc kubenswrapper[4967]: I1008 00:07:09.979826 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:09 crc kubenswrapper[4967]: I1008 00:07:09.979842 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:09 crc kubenswrapper[4967]: I1008 00:07:09.979854 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:09Z","lastTransitionTime":"2025-10-08T00:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:10 crc kubenswrapper[4967]: I1008 00:07:10.083159 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:10 crc kubenswrapper[4967]: I1008 00:07:10.083220 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:10 crc kubenswrapper[4967]: I1008 00:07:10.083239 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:10 crc kubenswrapper[4967]: I1008 00:07:10.083264 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:10 crc kubenswrapper[4967]: I1008 00:07:10.083283 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:10Z","lastTransitionTime":"2025-10-08T00:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:10 crc kubenswrapper[4967]: I1008 00:07:10.117865 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 00:07:10 crc kubenswrapper[4967]: E1008 00:07:10.118061 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 00:07:10 crc kubenswrapper[4967]: I1008 00:07:10.185421 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:10 crc kubenswrapper[4967]: I1008 00:07:10.185455 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:10 crc kubenswrapper[4967]: I1008 00:07:10.185465 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:10 crc kubenswrapper[4967]: I1008 00:07:10.185479 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:10 crc kubenswrapper[4967]: I1008 00:07:10.185492 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:10Z","lastTransitionTime":"2025-10-08T00:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:10 crc kubenswrapper[4967]: I1008 00:07:10.288149 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:10 crc kubenswrapper[4967]: I1008 00:07:10.288187 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:10 crc kubenswrapper[4967]: I1008 00:07:10.288199 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:10 crc kubenswrapper[4967]: I1008 00:07:10.288215 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:10 crc kubenswrapper[4967]: I1008 00:07:10.288225 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:10Z","lastTransitionTime":"2025-10-08T00:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:10 crc kubenswrapper[4967]: I1008 00:07:10.391506 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:10 crc kubenswrapper[4967]: I1008 00:07:10.391548 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:10 crc kubenswrapper[4967]: I1008 00:07:10.391573 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:10 crc kubenswrapper[4967]: I1008 00:07:10.391592 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:10 crc kubenswrapper[4967]: I1008 00:07:10.391606 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:10Z","lastTransitionTime":"2025-10-08T00:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:10 crc kubenswrapper[4967]: I1008 00:07:10.397057 4967 generic.go:334] "Generic (PLEG): container finished" podID="9299bc4f-f8d3-4baf-9885-39573982fd64" containerID="49f098966bfad49d06278eaa72be042596d488115e597737d9a0a0cfc78f99f1" exitCode=0 Oct 08 00:07:10 crc kubenswrapper[4967]: I1008 00:07:10.397103 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9mrr7" event={"ID":"9299bc4f-f8d3-4baf-9885-39573982fd64","Type":"ContainerDied","Data":"49f098966bfad49d06278eaa72be042596d488115e597737d9a0a0cfc78f99f1"} Oct 08 00:07:10 crc kubenswrapper[4967]: I1008 00:07:10.422241 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58c4704-a88a-43b4-b944-4de626f13f7e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22f5dabb39da0a7bd01db4dae1437b9a231a1fcc0dfe749a4a5a637a72cf33c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01a2b3dc785520c79eae998482e5a1347725c76584357073238fe60a619edb77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6aa6f5e81feceab3d1c7587152a70546b365e629a69267e4d3084324a1de764a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63c2b7c4b2791bd1f16888c58b878dbff1f87884e68c9949de914d0d3e1d39fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:10Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:10 crc kubenswrapper[4967]: I1008 00:07:10.448207 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:10Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:10 crc kubenswrapper[4967]: I1008 00:07:10.467935 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9mrr7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9299bc4f-f8d3-4baf-9885-39573982fd64\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb49c8476b482416588c4d403016940730064a619c0641eb029c947a4bab1215\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb49c8476b482416588c4d403016940730064a619c0641eb029c947a4bab1215\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b7b63962dcbf477f1fa4d71826a4aa3358920a2c1b2f5e48716411b5bdca049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b7b63962dcbf477f1fa4d71826a4aa3358920a2c1b2f5e48716411b5bdca049\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1e770dc737173fd4f3220ed6c52b422ccbd829722386a37029e1c3e10d80e1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1e770dc737173fd4f3220ed6c52b422ccbd829722386a37029e1c3e10d80e1a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e357d970878c996f219a79f1876b363176c9847b9575f33090f62bc299529a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e357d970878c996f219a79f1876b363176c9847b9575f33090f62bc299529a95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f098966bfad49d06278eaa72be042596d488115e597737d9a0a0cfc78f99f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49f098966bfad49d06278eaa72be042596d488115e597737d9a0a0cfc78f99f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9mrr7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:10Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:10 crc kubenswrapper[4967]: I1008 00:07:10.483608 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fa4b747-ab6e-4a11-9903-ef2dcb9668c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbadcdade274b6d9be940f22bc949e3c55cadf95dcad43c8cb7a1085c0e6708c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d0a0b2ef4a98ee828688081a74f4b7ecc31c8fae415d7257a560ecd43960ce3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96b53b1b283025b3d078a3f505d469d1872ad4bb8ab8e612c36c70b0b00a920\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1400e110997c64fa33d6b8672c7e91cab5f782952ddf8e1dea2a882689026e7f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1400e110997c64fa33d6b8672c7e91cab5f782952ddf8e1dea2a882689026e7f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 00:06:56.760088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 00:06:56.761210 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-369805205/tls.crt::/tmp/serving-cert-369805205/tls.key\\\\\\\"\\\\nI1008 00:07:02.321448 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 00:07:02.325932 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 00:07:02.325980 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 00:07:02.326025 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 00:07:02.326042 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 00:07:02.336203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 00:07:02.336240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 00:07:02.336245 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 00:07:02.336250 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 00:07:02.336253 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 00:07:02.336256 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 00:07:02.336258 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 00:07:02.336430 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 00:07:02.340774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:46Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca28d8db19036e47a7ef3dadb70941f6c6cf5735e7703977d036185c46fce40\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5b222ad84c31d8746dfe03124c7287b401deb3ca2c62ce2b4fbd40e638cee0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5b222ad84c31d8746dfe03124c7287b401deb3ca2c62ce2b4fbd40e638cee0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:10Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:10 crc kubenswrapper[4967]: I1008 00:07:10.496971 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:10 crc kubenswrapper[4967]: I1008 00:07:10.497004 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:10 crc kubenswrapper[4967]: I1008 00:07:10.497015 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:10 crc kubenswrapper[4967]: I1008 00:07:10.497034 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:10 crc kubenswrapper[4967]: I1008 00:07:10.497046 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:10Z","lastTransitionTime":"2025-10-08T00:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:10 crc kubenswrapper[4967]: I1008 00:07:10.498369 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:10Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:10 crc kubenswrapper[4967]: I1008 00:07:10.518983 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbfcdc882e2be1cb9cf666c9e7775671078f7dbbfe8421dd15db737b89920a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:10Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:10 crc kubenswrapper[4967]: I1008 00:07:10.536403 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9461162a-c27d-465e-9ca7-2d3936877cd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9778d78ddc7325abd4d77bc5e71b0b114f50be5dda4a438c7cb6b05640c8a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z9tp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28760af04ef321d0f91b2bce3ad074d44b63dabf676cc37f3be6aa8726341a2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z9tp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xvdtp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:10Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:10 crc kubenswrapper[4967]: I1008 00:07:10.549409 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94e63bc5641ce80d507f76a19b203f78c42368a7a4105f3fd09e19af21a1b735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:10Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:10 crc kubenswrapper[4967]: I1008 00:07:10.562106 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e98666fd3152bab57d3723e7da72aae2703b79f7ca73eb3c3e5f0e062a82592\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cf1cbe5548146fe359aafa351864125ec6b8bdd7b796aa4d311f6f3a16ad212\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:10Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:10 crc kubenswrapper[4967]: I1008 00:07:10.574323 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9tdbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41264a12-9486-4fee-a08c-607b590e5a70\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://806e85df796ae300d279f73a0cc4433a5bb1f7620ee8ce33c8dc4674c4ce0596\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5blv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9tdbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:10Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:10 crc kubenswrapper[4967]: I1008 00:07:10.592162 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f1af2ba-8907-4d14-afa1-ffde207c80f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bs8w9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:10Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:10 crc kubenswrapper[4967]: I1008 00:07:10.601531 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:10 crc kubenswrapper[4967]: I1008 00:07:10.601576 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:10 crc kubenswrapper[4967]: I1008 00:07:10.601590 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:10 crc kubenswrapper[4967]: I1008 00:07:10.601607 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:10 crc kubenswrapper[4967]: I1008 00:07:10.601620 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:10Z","lastTransitionTime":"2025-10-08T00:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:10 crc kubenswrapper[4967]: I1008 00:07:10.602887 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-v5n7t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abe7ae2d-7e63-4249-96e1-4000e9628ccc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52575e853ff6727dde7c89e4add4362be126127db27191c2a7bf2871efa719d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpsjd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:06Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-v5n7t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:10Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:10 crc kubenswrapper[4967]: I1008 00:07:10.614479 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:10Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:10 crc kubenswrapper[4967]: I1008 00:07:10.635092 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-npz2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c71501b1-aa9e-420e-b1bd-7e15365e7adc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dfe6c420d804b2d7daef4b5cfe3f87b77b369287c4d65553fa93f5f24f589d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gblk2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-npz2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:10Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:10 crc kubenswrapper[4967]: I1008 00:07:10.708118 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:10 crc kubenswrapper[4967]: I1008 00:07:10.708180 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:10 crc kubenswrapper[4967]: I1008 00:07:10.708197 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:10 crc kubenswrapper[4967]: I1008 00:07:10.708220 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:10 crc kubenswrapper[4967]: I1008 00:07:10.708236 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:10Z","lastTransitionTime":"2025-10-08T00:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:10 crc kubenswrapper[4967]: I1008 00:07:10.810598 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:10 crc kubenswrapper[4967]: I1008 00:07:10.810637 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:10 crc kubenswrapper[4967]: I1008 00:07:10.810647 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:10 crc kubenswrapper[4967]: I1008 00:07:10.810680 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:10 crc kubenswrapper[4967]: I1008 00:07:10.810690 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:10Z","lastTransitionTime":"2025-10-08T00:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:10 crc kubenswrapper[4967]: I1008 00:07:10.913443 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:10 crc kubenswrapper[4967]: I1008 00:07:10.913484 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:10 crc kubenswrapper[4967]: I1008 00:07:10.913497 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:10 crc kubenswrapper[4967]: I1008 00:07:10.913514 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:10 crc kubenswrapper[4967]: I1008 00:07:10.913526 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:10Z","lastTransitionTime":"2025-10-08T00:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:10 crc kubenswrapper[4967]: I1008 00:07:10.941585 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 00:07:10 crc kubenswrapper[4967]: I1008 00:07:10.941725 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 00:07:10 crc kubenswrapper[4967]: I1008 00:07:10.941762 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 00:07:10 crc kubenswrapper[4967]: I1008 00:07:10.941792 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 00:07:10 crc kubenswrapper[4967]: I1008 00:07:10.941815 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 00:07:10 crc kubenswrapper[4967]: E1008 00:07:10.941956 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 00:07:10 crc kubenswrapper[4967]: E1008 00:07:10.941974 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 00:07:10 crc kubenswrapper[4967]: E1008 00:07:10.941988 4967 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 00:07:10 crc kubenswrapper[4967]: E1008 00:07:10.942058 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-08 00:07:18.942042603 +0000 UTC m=+37.592577272 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 00:07:10 crc kubenswrapper[4967]: E1008 00:07:10.942159 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 00:07:18.942151256 +0000 UTC m=+37.592685915 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:07:10 crc kubenswrapper[4967]: E1008 00:07:10.942200 4967 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 00:07:10 crc kubenswrapper[4967]: E1008 00:07:10.942227 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 00:07:18.942219139 +0000 UTC m=+37.592753808 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 00:07:10 crc kubenswrapper[4967]: E1008 00:07:10.942348 4967 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 00:07:10 crc kubenswrapper[4967]: E1008 00:07:10.942437 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 00:07:18.942416454 +0000 UTC m=+37.592951113 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 00:07:10 crc kubenswrapper[4967]: E1008 00:07:10.942559 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 00:07:10 crc kubenswrapper[4967]: E1008 00:07:10.942575 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 00:07:10 crc kubenswrapper[4967]: E1008 00:07:10.942589 4967 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 00:07:10 crc kubenswrapper[4967]: E1008 00:07:10.942622 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-08 00:07:18.942611469 +0000 UTC m=+37.593146128 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.015472 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.015543 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.015557 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.015578 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.015592 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:11Z","lastTransitionTime":"2025-10-08T00:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.117050 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.117062 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 00:07:11 crc kubenswrapper[4967]: E1008 00:07:11.117262 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 00:07:11 crc kubenswrapper[4967]: E1008 00:07:11.117381 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.118239 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.118312 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.118332 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.118360 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.118380 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:11Z","lastTransitionTime":"2025-10-08T00:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.221447 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.221515 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.221534 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.221560 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.221578 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:11Z","lastTransitionTime":"2025-10-08T00:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.324939 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.324997 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.325016 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.325041 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.325060 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:11Z","lastTransitionTime":"2025-10-08T00:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.408909 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" event={"ID":"4f1af2ba-8907-4d14-afa1-ffde207c80f4","Type":"ContainerStarted","Data":"ead9cad55ea37f9a14c0beddb89065098e821e6b43db7e5ddacd89a2a12bf36a"} Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.409270 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.416889 4967 generic.go:334] "Generic (PLEG): container finished" podID="9299bc4f-f8d3-4baf-9885-39573982fd64" containerID="c1d53876b7b863c612d597fd716c3260e8857b6542802073bb730f6756071b46" exitCode=0 Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.416971 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9mrr7" event={"ID":"9299bc4f-f8d3-4baf-9885-39573982fd64","Type":"ContainerDied","Data":"c1d53876b7b863c612d597fd716c3260e8857b6542802073bb730f6756071b46"} Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.428181 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.428262 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.428291 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.428320 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.428341 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:11Z","lastTransitionTime":"2025-10-08T00:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.434570 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:11Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.454048 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.461372 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-npz2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c71501b1-aa9e-420e-b1bd-7e15365e7adc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dfe6c420d804b2d7daef4b5cfe3f87b77b369287c4d65553fa93f5f24f589d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gblk2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-npz2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:11Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.484019 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9mrr7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9299bc4f-f8d3-4baf-9885-39573982fd64\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb49c8476b482416588c4d403016940730064a619c0641eb029c947a4bab1215\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb49c8476b482416588c4d403016940730064a619c0641eb029c947a4bab1215\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b7b63962dcbf477f1fa4d71826a4aa3358920a2c1b2f5e48716411b5bdca049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b7b63962dcbf477f1fa4d71826a4aa3358920a2c1b2f5e48716411b5bdca049\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1e770dc737173fd4f3220ed6c52b422ccbd829722386a37029e1c3e10d80e1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1e770dc737173fd4f3220ed6c52b422ccbd829722386a37029e1c3e10d80e1a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e357d970878c996f219a79f1876b363176c9847b9575f33090f62bc299529a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e357d970878c996f219a79f1876b363176c9847b9575f33090f62bc299529a95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f098966bfad49d06278eaa72be042596d488115e597737d9a0a0cfc78f99f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49f098966bfad49d06278eaa72be042596d488115e597737d9a0a0cfc78f99f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9mrr7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:11Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.513365 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58c4704-a88a-43b4-b944-4de626f13f7e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22f5dabb39da0a7bd01db4dae1437b9a231a1fcc0dfe749a4a5a637a72cf33c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01a2b3dc785520c79eae998482e5a1347725c76584357073238fe60a619edb77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6aa6f5e81feceab3d1c7587152a70546b365e629a69267e4d3084324a1de764a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63c2b7c4b2791bd1f16888c58b878dbff1f87884e68c9949de914d0d3e1d39fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:11Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.529352 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:11Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.531152 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.531213 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.531226 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.531264 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.531276 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:11Z","lastTransitionTime":"2025-10-08T00:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.545116 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbfcdc882e2be1cb9cf666c9e7775671078f7dbbfe8421dd15db737b89920a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:11Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.556653 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9461162a-c27d-465e-9ca7-2d3936877cd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9778d78ddc7325abd4d77bc5e71b0b114f50be5dda4a438c7cb6b05640c8a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z9tp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28760af04ef321d0f91b2bce3ad074d44b63dabf676cc37f3be6aa8726341a2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z9tp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xvdtp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:11Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.570400 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fa4b747-ab6e-4a11-9903-ef2dcb9668c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbadcdade274b6d9be940f22bc949e3c55cadf95dcad43c8cb7a1085c0e6708c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d0a0b2ef4a98ee828688081a74f4b7ecc31c8fae415d7257a560ecd43960ce3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96b53b1b283025b3d078a3f505d469d1872ad4bb8ab8e612c36c70b0b00a920\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1400e110997c64fa33d6b8672c7e91cab5f782952ddf8e1dea2a882689026e7f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1400e110997c64fa33d6b8672c7e91cab5f782952ddf8e1dea2a882689026e7f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 00:06:56.760088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 00:06:56.761210 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-369805205/tls.crt::/tmp/serving-cert-369805205/tls.key\\\\\\\"\\\\nI1008 00:07:02.321448 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 00:07:02.325932 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 00:07:02.325980 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 00:07:02.326025 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 00:07:02.326042 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 00:07:02.336203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 00:07:02.336240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 00:07:02.336245 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 00:07:02.336250 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 00:07:02.336253 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 00:07:02.336256 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 00:07:02.336258 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 00:07:02.336430 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 00:07:02.340774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:46Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca28d8db19036e47a7ef3dadb70941f6c6cf5735e7703977d036185c46fce40\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5b222ad84c31d8746dfe03124c7287b401deb3ca2c62ce2b4fbd40e638cee0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5b222ad84c31d8746dfe03124c7287b401deb3ca2c62ce2b4fbd40e638cee0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:11Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.581852 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:11Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.597723 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f1af2ba-8907-4d14-afa1-ffde207c80f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4535cffb8c9d64e662e282ecb44f6817b4604866f0bb75015f3b888d5554f821\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b03b83ddff429dbfa8fe052901aaa86a39ff0dd75477ffa8a6b6f4198f5819ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c9fa0559eb35ba67e11e36069ac8d0b2fec76399788eef668a08ad4353df3ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3312263e12bd42fbed0476ab5ace66b5497f62c9477865e130d67e3922e05cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e69dc603055aa221d44d11c049ddcc50cec568979911a310ed773a90341a499\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115f7a33f3dedae8f59d2f45f465f95f48b99c824e895ef7a0be9fe7a1b61d23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ead9cad55ea37f9a14c0beddb89065098e821e6b43db7e5ddacd89a2a12bf36a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27fb2b1225e386dcd7fecaaa0aa54ba1fdfafabfde6c3824ae07315313067542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bs8w9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:11Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.609624 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-v5n7t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abe7ae2d-7e63-4249-96e1-4000e9628ccc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52575e853ff6727dde7c89e4add4362be126127db27191c2a7bf2871efa719d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpsjd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:06Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-v5n7t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:11Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.624579 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94e63bc5641ce80d507f76a19b203f78c42368a7a4105f3fd09e19af21a1b735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:11Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.633888 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.633958 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.633972 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.634067 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.634089 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:11Z","lastTransitionTime":"2025-10-08T00:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.637693 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e98666fd3152bab57d3723e7da72aae2703b79f7ca73eb3c3e5f0e062a82592\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cf1cbe5548146fe359aafa351864125ec6b8bdd7b796aa4d311f6f3a16ad212\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:11Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.648136 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9tdbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41264a12-9486-4fee-a08c-607b590e5a70\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://806e85df796ae300d279f73a0cc4433a5bb1f7620ee8ce33c8dc4674c4ce0596\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5blv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9tdbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:11Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.662000 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fa4b747-ab6e-4a11-9903-ef2dcb9668c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbadcdade274b6d9be940f22bc949e3c55cadf95dcad43c8cb7a1085c0e6708c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d0a0b2ef4a98ee828688081a74f4b7ecc31c8fae415d7257a560ecd43960ce3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96b53b1b283025b3d078a3f505d469d1872ad4bb8ab8e612c36c70b0b00a920\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1400e110997c64fa33d6b8672c7e91cab5f782952ddf8e1dea2a882689026e7f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1400e110997c64fa33d6b8672c7e91cab5f782952ddf8e1dea2a882689026e7f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 00:06:56.760088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 00:06:56.761210 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-369805205/tls.crt::/tmp/serving-cert-369805205/tls.key\\\\\\\"\\\\nI1008 00:07:02.321448 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 00:07:02.325932 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 00:07:02.325980 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 00:07:02.326025 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 00:07:02.326042 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 00:07:02.336203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 00:07:02.336240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 00:07:02.336245 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 00:07:02.336250 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 00:07:02.336253 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 00:07:02.336256 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 00:07:02.336258 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 00:07:02.336430 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 00:07:02.340774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:46Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca28d8db19036e47a7ef3dadb70941f6c6cf5735e7703977d036185c46fce40\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5b222ad84c31d8746dfe03124c7287b401deb3ca2c62ce2b4fbd40e638cee0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5b222ad84c31d8746dfe03124c7287b401deb3ca2c62ce2b4fbd40e638cee0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:11Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.673472 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:11Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.683336 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbfcdc882e2be1cb9cf666c9e7775671078f7dbbfe8421dd15db737b89920a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:11Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.695128 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9461162a-c27d-465e-9ca7-2d3936877cd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9778d78ddc7325abd4d77bc5e71b0b114f50be5dda4a438c7cb6b05640c8a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z9tp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28760af04ef321d0f91b2bce3ad074d44b63dabf676cc37f3be6aa8726341a2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z9tp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xvdtp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:11Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.706801 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94e63bc5641ce80d507f76a19b203f78c42368a7a4105f3fd09e19af21a1b735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:11Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.717476 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e98666fd3152bab57d3723e7da72aae2703b79f7ca73eb3c3e5f0e062a82592\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cf1cbe5548146fe359aafa351864125ec6b8bdd7b796aa4d311f6f3a16ad212\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:11Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.725521 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9tdbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41264a12-9486-4fee-a08c-607b590e5a70\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://806e85df796ae300d279f73a0cc4433a5bb1f7620ee8ce33c8dc4674c4ce0596\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5blv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9tdbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:11Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.736360 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.736404 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.736417 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.736434 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.736447 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:11Z","lastTransitionTime":"2025-10-08T00:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.741185 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f1af2ba-8907-4d14-afa1-ffde207c80f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4535cffb8c9d64e662e282ecb44f6817b4604866f0bb75015f3b888d5554f821\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b03b83ddff429dbfa8fe052901aaa86a39ff0dd75477ffa8a6b6f4198f5819ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c9fa0559eb35ba67e11e36069ac8d0b2fec76399788eef668a08ad4353df3ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3312263e12bd42fbed0476ab5ace66b5497f62c9477865e130d67e3922e05cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e69dc603055aa221d44d11c049ddcc50cec568979911a310ed773a90341a499\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115f7a33f3dedae8f59d2f45f465f95f48b99c824e895ef7a0be9fe7a1b61d23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ead9cad55ea37f9a14c0beddb89065098e821e6b43db7e5ddacd89a2a12bf36a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27fb2b1225e386dcd7fecaaa0aa54ba1fdfafabfde6c3824ae07315313067542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bs8w9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:11Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.749763 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-v5n7t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abe7ae2d-7e63-4249-96e1-4000e9628ccc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52575e853ff6727dde7c89e4add4362be126127db27191c2a7bf2871efa719d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpsjd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:06Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-v5n7t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:11Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.759857 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:11Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.772489 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-npz2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c71501b1-aa9e-420e-b1bd-7e15365e7adc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dfe6c420d804b2d7daef4b5cfe3f87b77b369287c4d65553fa93f5f24f589d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gblk2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-npz2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:11Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.783773 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58c4704-a88a-43b4-b944-4de626f13f7e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22f5dabb39da0a7bd01db4dae1437b9a231a1fcc0dfe749a4a5a637a72cf33c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01a2b3dc785520c79eae998482e5a1347725c76584357073238fe60a619edb77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6aa6f5e81feceab3d1c7587152a70546b365e629a69267e4d3084324a1de764a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63c2b7c4b2791bd1f16888c58b878dbff1f87884e68c9949de914d0d3e1d39fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:11Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.797972 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:11Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.815376 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9mrr7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9299bc4f-f8d3-4baf-9885-39573982fd64\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb49c8476b482416588c4d403016940730064a619c0641eb029c947a4bab1215\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb49c8476b482416588c4d403016940730064a619c0641eb029c947a4bab1215\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b7b63962dcbf477f1fa4d71826a4aa3358920a2c1b2f5e48716411b5bdca049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b7b63962dcbf477f1fa4d71826a4aa3358920a2c1b2f5e48716411b5bdca049\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1e770dc737173fd4f3220ed6c52b422ccbd829722386a37029e1c3e10d80e1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1e770dc737173fd4f3220ed6c52b422ccbd829722386a37029e1c3e10d80e1a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e357d970878c996f219a79f1876b363176c9847b9575f33090f62bc299529a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e357d970878c996f219a79f1876b363176c9847b9575f33090f62bc299529a95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f098966bfad49d06278eaa72be042596d488115e597737d9a0a0cfc78f99f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49f098966bfad49d06278eaa72be042596d488115e597737d9a0a0cfc78f99f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1d53876b7b863c612d597fd716c3260e8857b6542802073bb730f6756071b46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c1d53876b7b863c612d597fd716c3260e8857b6542802073bb730f6756071b46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9mrr7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:11Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.839040 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.839085 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.839097 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.839116 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.839126 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:11Z","lastTransitionTime":"2025-10-08T00:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.941830 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.941911 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.941931 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.941957 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:11 crc kubenswrapper[4967]: I1008 00:07:11.941976 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:11Z","lastTransitionTime":"2025-10-08T00:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.045054 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.045125 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.045144 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.045169 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.045190 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:12Z","lastTransitionTime":"2025-10-08T00:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.117956 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 00:07:12 crc kubenswrapper[4967]: E1008 00:07:12.118236 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.139189 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58c4704-a88a-43b4-b944-4de626f13f7e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22f5dabb39da0a7bd01db4dae1437b9a231a1fcc0dfe749a4a5a637a72cf33c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01a2b3dc785520c79eae998482e5a1347725c76584357073238fe60a619edb77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6aa6f5e81feceab3d1c7587152a70546b365e629a69267e4d3084324a1de764a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63c2b7c4b2791bd1f16888c58b878dbff1f87884e68c9949de914d0d3e1d39fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:12Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.147733 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.147788 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.147806 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.147832 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.147850 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:12Z","lastTransitionTime":"2025-10-08T00:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.163136 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:12Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.191189 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9mrr7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9299bc4f-f8d3-4baf-9885-39573982fd64\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb49c8476b482416588c4d403016940730064a619c0641eb029c947a4bab1215\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb49c8476b482416588c4d403016940730064a619c0641eb029c947a4bab1215\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b7b63962dcbf477f1fa4d71826a4aa3358920a2c1b2f5e48716411b5bdca049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b7b63962dcbf477f1fa4d71826a4aa3358920a2c1b2f5e48716411b5bdca049\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1e770dc737173fd4f3220ed6c52b422ccbd829722386a37029e1c3e10d80e1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1e770dc737173fd4f3220ed6c52b422ccbd829722386a37029e1c3e10d80e1a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e357d970878c996f219a79f1876b363176c9847b9575f33090f62bc299529a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e357d970878c996f219a79f1876b363176c9847b9575f33090f62bc299529a95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f098966bfad49d06278eaa72be042596d488115e597737d9a0a0cfc78f99f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49f098966bfad49d06278eaa72be042596d488115e597737d9a0a0cfc78f99f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1d53876b7b863c612d597fd716c3260e8857b6542802073bb730f6756071b46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c1d53876b7b863c612d597fd716c3260e8857b6542802073bb730f6756071b46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9mrr7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:12Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.216145 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fa4b747-ab6e-4a11-9903-ef2dcb9668c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbadcdade274b6d9be940f22bc949e3c55cadf95dcad43c8cb7a1085c0e6708c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d0a0b2ef4a98ee828688081a74f4b7ecc31c8fae415d7257a560ecd43960ce3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96b53b1b283025b3d078a3f505d469d1872ad4bb8ab8e612c36c70b0b00a920\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1400e110997c64fa33d6b8672c7e91cab5f782952ddf8e1dea2a882689026e7f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1400e110997c64fa33d6b8672c7e91cab5f782952ddf8e1dea2a882689026e7f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 00:06:56.760088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 00:06:56.761210 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-369805205/tls.crt::/tmp/serving-cert-369805205/tls.key\\\\\\\"\\\\nI1008 00:07:02.321448 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 00:07:02.325932 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 00:07:02.325980 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 00:07:02.326025 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 00:07:02.326042 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 00:07:02.336203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 00:07:02.336240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 00:07:02.336245 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 00:07:02.336250 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 00:07:02.336253 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 00:07:02.336256 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 00:07:02.336258 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 00:07:02.336430 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 00:07:02.340774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:46Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca28d8db19036e47a7ef3dadb70941f6c6cf5735e7703977d036185c46fce40\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5b222ad84c31d8746dfe03124c7287b401deb3ca2c62ce2b4fbd40e638cee0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5b222ad84c31d8746dfe03124c7287b401deb3ca2c62ce2b4fbd40e638cee0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:12Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.230346 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:12Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.242524 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbfcdc882e2be1cb9cf666c9e7775671078f7dbbfe8421dd15db737b89920a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:12Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.249653 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.249750 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.249762 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.249782 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.249797 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:12Z","lastTransitionTime":"2025-10-08T00:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.255150 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9461162a-c27d-465e-9ca7-2d3936877cd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9778d78ddc7325abd4d77bc5e71b0b114f50be5dda4a438c7cb6b05640c8a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z9tp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28760af04ef321d0f91b2bce3ad074d44b63dabf676cc37f3be6aa8726341a2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z9tp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xvdtp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:12Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.270516 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94e63bc5641ce80d507f76a19b203f78c42368a7a4105f3fd09e19af21a1b735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:12Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.288521 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e98666fd3152bab57d3723e7da72aae2703b79f7ca73eb3c3e5f0e062a82592\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cf1cbe5548146fe359aafa351864125ec6b8bdd7b796aa4d311f6f3a16ad212\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:12Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.300701 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9tdbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41264a12-9486-4fee-a08c-607b590e5a70\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://806e85df796ae300d279f73a0cc4433a5bb1f7620ee8ce33c8dc4674c4ce0596\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5blv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9tdbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:12Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.321345 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f1af2ba-8907-4d14-afa1-ffde207c80f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4535cffb8c9d64e662e282ecb44f6817b4604866f0bb75015f3b888d5554f821\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b03b83ddff429dbfa8fe052901aaa86a39ff0dd75477ffa8a6b6f4198f5819ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c9fa0559eb35ba67e11e36069ac8d0b2fec76399788eef668a08ad4353df3ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3312263e12bd42fbed0476ab5ace66b5497f62c9477865e130d67e3922e05cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e69dc603055aa221d44d11c049ddcc50cec568979911a310ed773a90341a499\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115f7a33f3dedae8f59d2f45f465f95f48b99c824e895ef7a0be9fe7a1b61d23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ead9cad55ea37f9a14c0beddb89065098e821e6b43db7e5ddacd89a2a12bf36a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27fb2b1225e386dcd7fecaaa0aa54ba1fdfafabfde6c3824ae07315313067542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bs8w9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:12Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.334878 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-v5n7t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abe7ae2d-7e63-4249-96e1-4000e9628ccc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52575e853ff6727dde7c89e4add4362be126127db27191c2a7bf2871efa719d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpsjd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:06Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-v5n7t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:12Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.350207 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:12Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.352373 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.352417 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.352434 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.352458 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.352476 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:12Z","lastTransitionTime":"2025-10-08T00:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.374297 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-npz2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c71501b1-aa9e-420e-b1bd-7e15365e7adc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dfe6c420d804b2d7daef4b5cfe3f87b77b369287c4d65553fa93f5f24f589d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gblk2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-npz2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:12Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.424559 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-9mrr7" event={"ID":"9299bc4f-f8d3-4baf-9885-39573982fd64","Type":"ContainerStarted","Data":"994beafece9dead0d85ae74dc453a142460705fd66b52ba22f2fa799cf27da08"} Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.424631 4967 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.425210 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.447515 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fa4b747-ab6e-4a11-9903-ef2dcb9668c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbadcdade274b6d9be940f22bc949e3c55cadf95dcad43c8cb7a1085c0e6708c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d0a0b2ef4a98ee828688081a74f4b7ecc31c8fae415d7257a560ecd43960ce3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96b53b1b283025b3d078a3f505d469d1872ad4bb8ab8e612c36c70b0b00a920\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1400e110997c64fa33d6b8672c7e91cab5f782952ddf8e1dea2a882689026e7f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1400e110997c64fa33d6b8672c7e91cab5f782952ddf8e1dea2a882689026e7f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 00:06:56.760088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 00:06:56.761210 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-369805205/tls.crt::/tmp/serving-cert-369805205/tls.key\\\\\\\"\\\\nI1008 00:07:02.321448 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 00:07:02.325932 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 00:07:02.325980 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 00:07:02.326025 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 00:07:02.326042 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 00:07:02.336203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 00:07:02.336240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 00:07:02.336245 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 00:07:02.336250 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 00:07:02.336253 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 00:07:02.336256 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 00:07:02.336258 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 00:07:02.336430 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 00:07:02.340774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:46Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca28d8db19036e47a7ef3dadb70941f6c6cf5735e7703977d036185c46fce40\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5b222ad84c31d8746dfe03124c7287b401deb3ca2c62ce2b4fbd40e638cee0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5b222ad84c31d8746dfe03124c7287b401deb3ca2c62ce2b4fbd40e638cee0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:12Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.454935 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.454989 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.455001 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.455017 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.455028 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:12Z","lastTransitionTime":"2025-10-08T00:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.458482 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.468977 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:12Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.481214 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbfcdc882e2be1cb9cf666c9e7775671078f7dbbfe8421dd15db737b89920a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:12Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.493432 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9461162a-c27d-465e-9ca7-2d3936877cd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9778d78ddc7325abd4d77bc5e71b0b114f50be5dda4a438c7cb6b05640c8a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z9tp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28760af04ef321d0f91b2bce3ad074d44b63dabf676cc37f3be6aa8726341a2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z9tp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xvdtp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:12Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.506959 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94e63bc5641ce80d507f76a19b203f78c42368a7a4105f3fd09e19af21a1b735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:12Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.517936 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e98666fd3152bab57d3723e7da72aae2703b79f7ca73eb3c3e5f0e062a82592\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cf1cbe5548146fe359aafa351864125ec6b8bdd7b796aa4d311f6f3a16ad212\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:12Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.532471 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9tdbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41264a12-9486-4fee-a08c-607b590e5a70\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://806e85df796ae300d279f73a0cc4433a5bb1f7620ee8ce33c8dc4674c4ce0596\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5blv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9tdbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:12Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.549417 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f1af2ba-8907-4d14-afa1-ffde207c80f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4535cffb8c9d64e662e282ecb44f6817b4604866f0bb75015f3b888d5554f821\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b03b83ddff429dbfa8fe052901aaa86a39ff0dd75477ffa8a6b6f4198f5819ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c9fa0559eb35ba67e11e36069ac8d0b2fec76399788eef668a08ad4353df3ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3312263e12bd42fbed0476ab5ace66b5497f62c9477865e130d67e3922e05cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e69dc603055aa221d44d11c049ddcc50cec568979911a310ed773a90341a499\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115f7a33f3dedae8f59d2f45f465f95f48b99c824e895ef7a0be9fe7a1b61d23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ead9cad55ea37f9a14c0beddb89065098e821e6b43db7e5ddacd89a2a12bf36a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27fb2b1225e386dcd7fecaaa0aa54ba1fdfafabfde6c3824ae07315313067542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bs8w9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:12Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.557790 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.557846 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.557864 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.557889 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.557907 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:12Z","lastTransitionTime":"2025-10-08T00:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.559034 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-v5n7t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abe7ae2d-7e63-4249-96e1-4000e9628ccc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52575e853ff6727dde7c89e4add4362be126127db27191c2a7bf2871efa719d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpsjd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:06Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-v5n7t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:12Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.573084 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:12Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.585773 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-npz2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c71501b1-aa9e-420e-b1bd-7e15365e7adc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dfe6c420d804b2d7daef4b5cfe3f87b77b369287c4d65553fa93f5f24f589d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gblk2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-npz2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:12Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.599188 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58c4704-a88a-43b4-b944-4de626f13f7e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22f5dabb39da0a7bd01db4dae1437b9a231a1fcc0dfe749a4a5a637a72cf33c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01a2b3dc785520c79eae998482e5a1347725c76584357073238fe60a619edb77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6aa6f5e81feceab3d1c7587152a70546b365e629a69267e4d3084324a1de764a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63c2b7c4b2791bd1f16888c58b878dbff1f87884e68c9949de914d0d3e1d39fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:12Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.614557 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:12Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.635076 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9mrr7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9299bc4f-f8d3-4baf-9885-39573982fd64\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://994beafece9dead0d85ae74dc453a142460705fd66b52ba22f2fa799cf27da08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb49c8476b482416588c4d403016940730064a619c0641eb029c947a4bab1215\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb49c8476b482416588c4d403016940730064a619c0641eb029c947a4bab1215\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b7b63962dcbf477f1fa4d71826a4aa3358920a2c1b2f5e48716411b5bdca049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b7b63962dcbf477f1fa4d71826a4aa3358920a2c1b2f5e48716411b5bdca049\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1e770dc737173fd4f3220ed6c52b422ccbd829722386a37029e1c3e10d80e1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1e770dc737173fd4f3220ed6c52b422ccbd829722386a37029e1c3e10d80e1a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e357d970878c996f219a79f1876b363176c9847b9575f33090f62bc299529a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e357d970878c996f219a79f1876b363176c9847b9575f33090f62bc299529a95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f098966bfad49d06278eaa72be042596d488115e597737d9a0a0cfc78f99f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49f098966bfad49d06278eaa72be042596d488115e597737d9a0a0cfc78f99f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1d53876b7b863c612d597fd716c3260e8857b6542802073bb730f6756071b46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c1d53876b7b863c612d597fd716c3260e8857b6542802073bb730f6756071b46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9mrr7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:12Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.652095 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94e63bc5641ce80d507f76a19b203f78c42368a7a4105f3fd09e19af21a1b735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:12Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.660641 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.660735 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.660760 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.660784 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.660801 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:12Z","lastTransitionTime":"2025-10-08T00:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.669031 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e98666fd3152bab57d3723e7da72aae2703b79f7ca73eb3c3e5f0e062a82592\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cf1cbe5548146fe359aafa351864125ec6b8bdd7b796aa4d311f6f3a16ad212\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:12Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.683599 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9tdbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41264a12-9486-4fee-a08c-607b590e5a70\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://806e85df796ae300d279f73a0cc4433a5bb1f7620ee8ce33c8dc4674c4ce0596\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5blv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9tdbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:12Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.706088 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f1af2ba-8907-4d14-afa1-ffde207c80f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4535cffb8c9d64e662e282ecb44f6817b4604866f0bb75015f3b888d5554f821\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b03b83ddff429dbfa8fe052901aaa86a39ff0dd75477ffa8a6b6f4198f5819ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c9fa0559eb35ba67e11e36069ac8d0b2fec76399788eef668a08ad4353df3ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3312263e12bd42fbed0476ab5ace66b5497f62c9477865e130d67e3922e05cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e69dc603055aa221d44d11c049ddcc50cec568979911a310ed773a90341a499\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115f7a33f3dedae8f59d2f45f465f95f48b99c824e895ef7a0be9fe7a1b61d23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ead9cad55ea37f9a14c0beddb89065098e821e6b43db7e5ddacd89a2a12bf36a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27fb2b1225e386dcd7fecaaa0aa54ba1fdfafabfde6c3824ae07315313067542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bs8w9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:12Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.717247 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-v5n7t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abe7ae2d-7e63-4249-96e1-4000e9628ccc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52575e853ff6727dde7c89e4add4362be126127db27191c2a7bf2871efa719d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpsjd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:06Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-v5n7t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:12Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.730522 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:12Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.750724 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-npz2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c71501b1-aa9e-420e-b1bd-7e15365e7adc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dfe6c420d804b2d7daef4b5cfe3f87b77b369287c4d65553fa93f5f24f589d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gblk2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-npz2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:12Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.765815 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.765926 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.765952 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.765996 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.766019 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:12Z","lastTransitionTime":"2025-10-08T00:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.777959 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58c4704-a88a-43b4-b944-4de626f13f7e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22f5dabb39da0a7bd01db4dae1437b9a231a1fcc0dfe749a4a5a637a72cf33c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01a2b3dc785520c79eae998482e5a1347725c76584357073238fe60a619edb77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6aa6f5e81feceab3d1c7587152a70546b365e629a69267e4d3084324a1de764a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63c2b7c4b2791bd1f16888c58b878dbff1f87884e68c9949de914d0d3e1d39fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:12Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.829524 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:12Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.855702 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9mrr7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9299bc4f-f8d3-4baf-9885-39573982fd64\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://994beafece9dead0d85ae74dc453a142460705fd66b52ba22f2fa799cf27da08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb49c8476b482416588c4d403016940730064a619c0641eb029c947a4bab1215\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb49c8476b482416588c4d403016940730064a619c0641eb029c947a4bab1215\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b7b63962dcbf477f1fa4d71826a4aa3358920a2c1b2f5e48716411b5bdca049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b7b63962dcbf477f1fa4d71826a4aa3358920a2c1b2f5e48716411b5bdca049\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1e770dc737173fd4f3220ed6c52b422ccbd829722386a37029e1c3e10d80e1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1e770dc737173fd4f3220ed6c52b422ccbd829722386a37029e1c3e10d80e1a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e357d970878c996f219a79f1876b363176c9847b9575f33090f62bc299529a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e357d970878c996f219a79f1876b363176c9847b9575f33090f62bc299529a95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f098966bfad49d06278eaa72be042596d488115e597737d9a0a0cfc78f99f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49f098966bfad49d06278eaa72be042596d488115e597737d9a0a0cfc78f99f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1d53876b7b863c612d597fd716c3260e8857b6542802073bb730f6756071b46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c1d53876b7b863c612d597fd716c3260e8857b6542802073bb730f6756071b46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9mrr7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:12Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.867959 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.867996 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.868006 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.868021 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.868030 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:12Z","lastTransitionTime":"2025-10-08T00:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.876559 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fa4b747-ab6e-4a11-9903-ef2dcb9668c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbadcdade274b6d9be940f22bc949e3c55cadf95dcad43c8cb7a1085c0e6708c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d0a0b2ef4a98ee828688081a74f4b7ecc31c8fae415d7257a560ecd43960ce3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96b53b1b283025b3d078a3f505d469d1872ad4bb8ab8e612c36c70b0b00a920\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1400e110997c64fa33d6b8672c7e91cab5f782952ddf8e1dea2a882689026e7f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1400e110997c64fa33d6b8672c7e91cab5f782952ddf8e1dea2a882689026e7f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 00:06:56.760088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 00:06:56.761210 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-369805205/tls.crt::/tmp/serving-cert-369805205/tls.key\\\\\\\"\\\\nI1008 00:07:02.321448 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 00:07:02.325932 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 00:07:02.325980 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 00:07:02.326025 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 00:07:02.326042 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 00:07:02.336203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 00:07:02.336240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 00:07:02.336245 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 00:07:02.336250 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 00:07:02.336253 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 00:07:02.336256 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 00:07:02.336258 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 00:07:02.336430 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 00:07:02.340774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:46Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca28d8db19036e47a7ef3dadb70941f6c6cf5735e7703977d036185c46fce40\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5b222ad84c31d8746dfe03124c7287b401deb3ca2c62ce2b4fbd40e638cee0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5b222ad84c31d8746dfe03124c7287b401deb3ca2c62ce2b4fbd40e638cee0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:12Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.893931 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:12Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.908762 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbfcdc882e2be1cb9cf666c9e7775671078f7dbbfe8421dd15db737b89920a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:12Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.927737 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9461162a-c27d-465e-9ca7-2d3936877cd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9778d78ddc7325abd4d77bc5e71b0b114f50be5dda4a438c7cb6b05640c8a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z9tp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28760af04ef321d0f91b2bce3ad074d44b63dabf676cc37f3be6aa8726341a2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z9tp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xvdtp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:12Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.970484 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.970517 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.970527 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.970542 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:12 crc kubenswrapper[4967]: I1008 00:07:12.970552 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:12Z","lastTransitionTime":"2025-10-08T00:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:13 crc kubenswrapper[4967]: I1008 00:07:13.072508 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:13 crc kubenswrapper[4967]: I1008 00:07:13.072548 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:13 crc kubenswrapper[4967]: I1008 00:07:13.072560 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:13 crc kubenswrapper[4967]: I1008 00:07:13.072576 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:13 crc kubenswrapper[4967]: I1008 00:07:13.072587 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:13Z","lastTransitionTime":"2025-10-08T00:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:13 crc kubenswrapper[4967]: I1008 00:07:13.117352 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 00:07:13 crc kubenswrapper[4967]: E1008 00:07:13.117463 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 00:07:13 crc kubenswrapper[4967]: I1008 00:07:13.117901 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 00:07:13 crc kubenswrapper[4967]: E1008 00:07:13.118057 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 00:07:13 crc kubenswrapper[4967]: I1008 00:07:13.175069 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:13 crc kubenswrapper[4967]: I1008 00:07:13.175104 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:13 crc kubenswrapper[4967]: I1008 00:07:13.175114 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:13 crc kubenswrapper[4967]: I1008 00:07:13.175129 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:13 crc kubenswrapper[4967]: I1008 00:07:13.175138 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:13Z","lastTransitionTime":"2025-10-08T00:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:13 crc kubenswrapper[4967]: I1008 00:07:13.277654 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:13 crc kubenswrapper[4967]: I1008 00:07:13.277740 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:13 crc kubenswrapper[4967]: I1008 00:07:13.277759 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:13 crc kubenswrapper[4967]: I1008 00:07:13.278046 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:13 crc kubenswrapper[4967]: I1008 00:07:13.278070 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:13Z","lastTransitionTime":"2025-10-08T00:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:13 crc kubenswrapper[4967]: I1008 00:07:13.392600 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:13 crc kubenswrapper[4967]: I1008 00:07:13.392635 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:13 crc kubenswrapper[4967]: I1008 00:07:13.392643 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:13 crc kubenswrapper[4967]: I1008 00:07:13.392656 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:13 crc kubenswrapper[4967]: I1008 00:07:13.392680 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:13Z","lastTransitionTime":"2025-10-08T00:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:13 crc kubenswrapper[4967]: I1008 00:07:13.427584 4967 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 08 00:07:13 crc kubenswrapper[4967]: I1008 00:07:13.494998 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:13 crc kubenswrapper[4967]: I1008 00:07:13.495044 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:13 crc kubenswrapper[4967]: I1008 00:07:13.495060 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:13 crc kubenswrapper[4967]: I1008 00:07:13.495079 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:13 crc kubenswrapper[4967]: I1008 00:07:13.495095 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:13Z","lastTransitionTime":"2025-10-08T00:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:13 crc kubenswrapper[4967]: I1008 00:07:13.602432 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:13 crc kubenswrapper[4967]: I1008 00:07:13.602492 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:13 crc kubenswrapper[4967]: I1008 00:07:13.602512 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:13 crc kubenswrapper[4967]: I1008 00:07:13.602537 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:13 crc kubenswrapper[4967]: I1008 00:07:13.602556 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:13Z","lastTransitionTime":"2025-10-08T00:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:13 crc kubenswrapper[4967]: I1008 00:07:13.717040 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:13 crc kubenswrapper[4967]: I1008 00:07:13.717105 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:13 crc kubenswrapper[4967]: I1008 00:07:13.717124 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:13 crc kubenswrapper[4967]: I1008 00:07:13.717158 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:13 crc kubenswrapper[4967]: I1008 00:07:13.717175 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:13Z","lastTransitionTime":"2025-10-08T00:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:13 crc kubenswrapper[4967]: I1008 00:07:13.819510 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:13 crc kubenswrapper[4967]: I1008 00:07:13.819593 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:13 crc kubenswrapper[4967]: I1008 00:07:13.819619 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:13 crc kubenswrapper[4967]: I1008 00:07:13.819657 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:13 crc kubenswrapper[4967]: I1008 00:07:13.819719 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:13Z","lastTransitionTime":"2025-10-08T00:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:13 crc kubenswrapper[4967]: I1008 00:07:13.922282 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:13 crc kubenswrapper[4967]: I1008 00:07:13.922354 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:13 crc kubenswrapper[4967]: I1008 00:07:13.922390 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:13 crc kubenswrapper[4967]: I1008 00:07:13.922424 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:13 crc kubenswrapper[4967]: I1008 00:07:13.922449 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:13Z","lastTransitionTime":"2025-10-08T00:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:14 crc kubenswrapper[4967]: I1008 00:07:14.024592 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:14 crc kubenswrapper[4967]: I1008 00:07:14.024629 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:14 crc kubenswrapper[4967]: I1008 00:07:14.024639 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:14 crc kubenswrapper[4967]: I1008 00:07:14.024654 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:14 crc kubenswrapper[4967]: I1008 00:07:14.024663 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:14Z","lastTransitionTime":"2025-10-08T00:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:14 crc kubenswrapper[4967]: I1008 00:07:14.117361 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 00:07:14 crc kubenswrapper[4967]: E1008 00:07:14.117479 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 00:07:14 crc kubenswrapper[4967]: I1008 00:07:14.126155 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:14 crc kubenswrapper[4967]: I1008 00:07:14.126203 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:14 crc kubenswrapper[4967]: I1008 00:07:14.126220 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:14 crc kubenswrapper[4967]: I1008 00:07:14.126240 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:14 crc kubenswrapper[4967]: I1008 00:07:14.126257 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:14Z","lastTransitionTime":"2025-10-08T00:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:14 crc kubenswrapper[4967]: I1008 00:07:14.228143 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:14 crc kubenswrapper[4967]: I1008 00:07:14.228183 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:14 crc kubenswrapper[4967]: I1008 00:07:14.228196 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:14 crc kubenswrapper[4967]: I1008 00:07:14.228213 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:14 crc kubenswrapper[4967]: I1008 00:07:14.228225 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:14Z","lastTransitionTime":"2025-10-08T00:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:14 crc kubenswrapper[4967]: I1008 00:07:14.331165 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:14 crc kubenswrapper[4967]: I1008 00:07:14.331206 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:14 crc kubenswrapper[4967]: I1008 00:07:14.331220 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:14 crc kubenswrapper[4967]: I1008 00:07:14.331235 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:14 crc kubenswrapper[4967]: I1008 00:07:14.331244 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:14Z","lastTransitionTime":"2025-10-08T00:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:14 crc kubenswrapper[4967]: I1008 00:07:14.432253 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bs8w9_4f1af2ba-8907-4d14-afa1-ffde207c80f4/ovnkube-controller/0.log" Oct 08 00:07:14 crc kubenswrapper[4967]: I1008 00:07:14.433336 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:14 crc kubenswrapper[4967]: I1008 00:07:14.433393 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:14 crc kubenswrapper[4967]: I1008 00:07:14.433409 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:14 crc kubenswrapper[4967]: I1008 00:07:14.433430 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:14 crc kubenswrapper[4967]: I1008 00:07:14.433446 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:14Z","lastTransitionTime":"2025-10-08T00:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:14 crc kubenswrapper[4967]: I1008 00:07:14.435110 4967 generic.go:334] "Generic (PLEG): container finished" podID="4f1af2ba-8907-4d14-afa1-ffde207c80f4" containerID="ead9cad55ea37f9a14c0beddb89065098e821e6b43db7e5ddacd89a2a12bf36a" exitCode=1 Oct 08 00:07:14 crc kubenswrapper[4967]: I1008 00:07:14.435208 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" event={"ID":"4f1af2ba-8907-4d14-afa1-ffde207c80f4","Type":"ContainerDied","Data":"ead9cad55ea37f9a14c0beddb89065098e821e6b43db7e5ddacd89a2a12bf36a"} Oct 08 00:07:14 crc kubenswrapper[4967]: I1008 00:07:14.435772 4967 scope.go:117] "RemoveContainer" containerID="ead9cad55ea37f9a14c0beddb89065098e821e6b43db7e5ddacd89a2a12bf36a" Oct 08 00:07:14 crc kubenswrapper[4967]: I1008 00:07:14.454990 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fa4b747-ab6e-4a11-9903-ef2dcb9668c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbadcdade274b6d9be940f22bc949e3c55cadf95dcad43c8cb7a1085c0e6708c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d0a0b2ef4a98ee828688081a74f4b7ecc31c8fae415d7257a560ecd43960ce3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96b53b1b283025b3d078a3f505d469d1872ad4bb8ab8e612c36c70b0b00a920\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1400e110997c64fa33d6b8672c7e91cab5f782952ddf8e1dea2a882689026e7f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1400e110997c64fa33d6b8672c7e91cab5f782952ddf8e1dea2a882689026e7f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 00:06:56.760088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 00:06:56.761210 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-369805205/tls.crt::/tmp/serving-cert-369805205/tls.key\\\\\\\"\\\\nI1008 00:07:02.321448 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 00:07:02.325932 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 00:07:02.325980 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 00:07:02.326025 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 00:07:02.326042 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 00:07:02.336203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 00:07:02.336240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 00:07:02.336245 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 00:07:02.336250 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 00:07:02.336253 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 00:07:02.336256 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 00:07:02.336258 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 00:07:02.336430 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 00:07:02.340774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:46Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca28d8db19036e47a7ef3dadb70941f6c6cf5735e7703977d036185c46fce40\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5b222ad84c31d8746dfe03124c7287b401deb3ca2c62ce2b4fbd40e638cee0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5b222ad84c31d8746dfe03124c7287b401deb3ca2c62ce2b4fbd40e638cee0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:14Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:14 crc kubenswrapper[4967]: I1008 00:07:14.475622 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:14Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:14 crc kubenswrapper[4967]: I1008 00:07:14.491724 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbfcdc882e2be1cb9cf666c9e7775671078f7dbbfe8421dd15db737b89920a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:14Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:14 crc kubenswrapper[4967]: I1008 00:07:14.510554 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9461162a-c27d-465e-9ca7-2d3936877cd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9778d78ddc7325abd4d77bc5e71b0b114f50be5dda4a438c7cb6b05640c8a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z9tp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28760af04ef321d0f91b2bce3ad074d44b63dabf676cc37f3be6aa8726341a2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z9tp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xvdtp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:14Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:14 crc kubenswrapper[4967]: I1008 00:07:14.529336 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94e63bc5641ce80d507f76a19b203f78c42368a7a4105f3fd09e19af21a1b735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:14Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:14 crc kubenswrapper[4967]: I1008 00:07:14.536324 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:14 crc kubenswrapper[4967]: I1008 00:07:14.536367 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:14 crc kubenswrapper[4967]: I1008 00:07:14.536380 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:14 crc kubenswrapper[4967]: I1008 00:07:14.536400 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:14 crc kubenswrapper[4967]: I1008 00:07:14.536412 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:14Z","lastTransitionTime":"2025-10-08T00:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:14 crc kubenswrapper[4967]: I1008 00:07:14.542821 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e98666fd3152bab57d3723e7da72aae2703b79f7ca73eb3c3e5f0e062a82592\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cf1cbe5548146fe359aafa351864125ec6b8bdd7b796aa4d311f6f3a16ad212\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:14Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:14 crc kubenswrapper[4967]: I1008 00:07:14.552997 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9tdbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41264a12-9486-4fee-a08c-607b590e5a70\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://806e85df796ae300d279f73a0cc4433a5bb1f7620ee8ce33c8dc4674c4ce0596\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5blv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9tdbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:14Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:14 crc kubenswrapper[4967]: I1008 00:07:14.576973 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f1af2ba-8907-4d14-afa1-ffde207c80f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4535cffb8c9d64e662e282ecb44f6817b4604866f0bb75015f3b888d5554f821\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b03b83ddff429dbfa8fe052901aaa86a39ff0dd75477ffa8a6b6f4198f5819ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c9fa0559eb35ba67e11e36069ac8d0b2fec76399788eef668a08ad4353df3ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3312263e12bd42fbed0476ab5ace66b5497f62c9477865e130d67e3922e05cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e69dc603055aa221d44d11c049ddcc50cec568979911a310ed773a90341a499\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115f7a33f3dedae8f59d2f45f465f95f48b99c824e895ef7a0be9fe7a1b61d23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ead9cad55ea37f9a14c0beddb89065098e821e6b43db7e5ddacd89a2a12bf36a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ead9cad55ea37f9a14c0beddb89065098e821e6b43db7e5ddacd89a2a12bf36a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T00:07:13Z\\\",\\\"message\\\":\\\"tory.go:140\\\\nI1008 00:07:13.601660 6243 reflector.go:311] Stopping reflector *v1.ClusterUserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1008 00:07:13.601605 6243 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1008 00:07:13.601819 6243 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1008 00:07:13.601862 6243 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1008 00:07:13.601897 6243 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1008 00:07:13.601932 6243 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1008 00:07:13.601973 6243 factory.go:656] Stopping watch factory\\\\nI1008 00:07:13.602003 6243 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1008 00:07:13.601622 6243 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1008 00:07:13.602415 6243 handler.go:208] Removed *v1.Node event handler 7\\\\nI1008 00:07:13.602437 6243 handler.go:208] Removed *v1.Node event handler 2\\\\nI1008 00:07:13.603125 6243 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1008 00:07:13.603186 6243 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1008 00:07:13.603406 6243 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27fb2b1225e386dcd7fecaaa0aa54ba1fdfafabfde6c3824ae07315313067542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bs8w9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:14Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:14 crc kubenswrapper[4967]: I1008 00:07:14.587743 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-v5n7t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abe7ae2d-7e63-4249-96e1-4000e9628ccc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52575e853ff6727dde7c89e4add4362be126127db27191c2a7bf2871efa719d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpsjd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:06Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-v5n7t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:14Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:14 crc kubenswrapper[4967]: I1008 00:07:14.604415 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:14Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:14 crc kubenswrapper[4967]: I1008 00:07:14.618951 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-npz2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c71501b1-aa9e-420e-b1bd-7e15365e7adc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dfe6c420d804b2d7daef4b5cfe3f87b77b369287c4d65553fa93f5f24f589d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gblk2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-npz2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:14Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:14 crc kubenswrapper[4967]: I1008 00:07:14.635930 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58c4704-a88a-43b4-b944-4de626f13f7e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22f5dabb39da0a7bd01db4dae1437b9a231a1fcc0dfe749a4a5a637a72cf33c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01a2b3dc785520c79eae998482e5a1347725c76584357073238fe60a619edb77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6aa6f5e81feceab3d1c7587152a70546b365e629a69267e4d3084324a1de764a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63c2b7c4b2791bd1f16888c58b878dbff1f87884e68c9949de914d0d3e1d39fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:14Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:14 crc kubenswrapper[4967]: I1008 00:07:14.638802 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:14 crc kubenswrapper[4967]: I1008 00:07:14.638831 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:14 crc kubenswrapper[4967]: I1008 00:07:14.638924 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:14 crc kubenswrapper[4967]: I1008 00:07:14.638940 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:14 crc kubenswrapper[4967]: I1008 00:07:14.638948 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:14Z","lastTransitionTime":"2025-10-08T00:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:14 crc kubenswrapper[4967]: I1008 00:07:14.648436 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:14Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:14 crc kubenswrapper[4967]: I1008 00:07:14.664716 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9mrr7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9299bc4f-f8d3-4baf-9885-39573982fd64\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://994beafece9dead0d85ae74dc453a142460705fd66b52ba22f2fa799cf27da08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb49c8476b482416588c4d403016940730064a619c0641eb029c947a4bab1215\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb49c8476b482416588c4d403016940730064a619c0641eb029c947a4bab1215\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b7b63962dcbf477f1fa4d71826a4aa3358920a2c1b2f5e48716411b5bdca049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b7b63962dcbf477f1fa4d71826a4aa3358920a2c1b2f5e48716411b5bdca049\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1e770dc737173fd4f3220ed6c52b422ccbd829722386a37029e1c3e10d80e1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1e770dc737173fd4f3220ed6c52b422ccbd829722386a37029e1c3e10d80e1a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e357d970878c996f219a79f1876b363176c9847b9575f33090f62bc299529a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e357d970878c996f219a79f1876b363176c9847b9575f33090f62bc299529a95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f098966bfad49d06278eaa72be042596d488115e597737d9a0a0cfc78f99f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49f098966bfad49d06278eaa72be042596d488115e597737d9a0a0cfc78f99f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1d53876b7b863c612d597fd716c3260e8857b6542802073bb730f6756071b46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c1d53876b7b863c612d597fd716c3260e8857b6542802073bb730f6756071b46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9mrr7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:14Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:14 crc kubenswrapper[4967]: I1008 00:07:14.741352 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:14 crc kubenswrapper[4967]: I1008 00:07:14.741386 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:14 crc kubenswrapper[4967]: I1008 00:07:14.741397 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:14 crc kubenswrapper[4967]: I1008 00:07:14.741413 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:14 crc kubenswrapper[4967]: I1008 00:07:14.741422 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:14Z","lastTransitionTime":"2025-10-08T00:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:14 crc kubenswrapper[4967]: I1008 00:07:14.844592 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:14 crc kubenswrapper[4967]: I1008 00:07:14.844641 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:14 crc kubenswrapper[4967]: I1008 00:07:14.844658 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:14 crc kubenswrapper[4967]: I1008 00:07:14.844720 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:14 crc kubenswrapper[4967]: I1008 00:07:14.844738 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:14Z","lastTransitionTime":"2025-10-08T00:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:14 crc kubenswrapper[4967]: I1008 00:07:14.947238 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:14 crc kubenswrapper[4967]: I1008 00:07:14.947274 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:14 crc kubenswrapper[4967]: I1008 00:07:14.947286 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:14 crc kubenswrapper[4967]: I1008 00:07:14.947349 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:14 crc kubenswrapper[4967]: I1008 00:07:14.947360 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:14Z","lastTransitionTime":"2025-10-08T00:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:15 crc kubenswrapper[4967]: I1008 00:07:15.050157 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:15 crc kubenswrapper[4967]: I1008 00:07:15.050199 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:15 crc kubenswrapper[4967]: I1008 00:07:15.050212 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:15 crc kubenswrapper[4967]: I1008 00:07:15.050230 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:15 crc kubenswrapper[4967]: I1008 00:07:15.050241 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:15Z","lastTransitionTime":"2025-10-08T00:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:15 crc kubenswrapper[4967]: I1008 00:07:15.117442 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 00:07:15 crc kubenswrapper[4967]: I1008 00:07:15.117500 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 00:07:15 crc kubenswrapper[4967]: E1008 00:07:15.117588 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 00:07:15 crc kubenswrapper[4967]: E1008 00:07:15.117747 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 00:07:15 crc kubenswrapper[4967]: I1008 00:07:15.152615 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:15 crc kubenswrapper[4967]: I1008 00:07:15.152657 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:15 crc kubenswrapper[4967]: I1008 00:07:15.152689 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:15 crc kubenswrapper[4967]: I1008 00:07:15.152716 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:15 crc kubenswrapper[4967]: I1008 00:07:15.152730 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:15Z","lastTransitionTime":"2025-10-08T00:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:15 crc kubenswrapper[4967]: I1008 00:07:15.254892 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:15 crc kubenswrapper[4967]: I1008 00:07:15.254930 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:15 crc kubenswrapper[4967]: I1008 00:07:15.254944 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:15 crc kubenswrapper[4967]: I1008 00:07:15.254959 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:15 crc kubenswrapper[4967]: I1008 00:07:15.254970 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:15Z","lastTransitionTime":"2025-10-08T00:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:15 crc kubenswrapper[4967]: I1008 00:07:15.357175 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:15 crc kubenswrapper[4967]: I1008 00:07:15.357209 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:15 crc kubenswrapper[4967]: I1008 00:07:15.357219 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:15 crc kubenswrapper[4967]: I1008 00:07:15.357231 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:15 crc kubenswrapper[4967]: I1008 00:07:15.357242 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:15Z","lastTransitionTime":"2025-10-08T00:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:15 crc kubenswrapper[4967]: I1008 00:07:15.440496 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bs8w9_4f1af2ba-8907-4d14-afa1-ffde207c80f4/ovnkube-controller/0.log" Oct 08 00:07:15 crc kubenswrapper[4967]: I1008 00:07:15.444100 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" event={"ID":"4f1af2ba-8907-4d14-afa1-ffde207c80f4","Type":"ContainerStarted","Data":"f9972551ade22d86d9ccd63097a793a2fb4ce74779a99f4b9aab1cf215087975"} Oct 08 00:07:15 crc kubenswrapper[4967]: I1008 00:07:15.444264 4967 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 08 00:07:15 crc kubenswrapper[4967]: I1008 00:07:15.459439 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:15 crc kubenswrapper[4967]: I1008 00:07:15.459529 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:15 crc kubenswrapper[4967]: I1008 00:07:15.459548 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:15 crc kubenswrapper[4967]: I1008 00:07:15.459571 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:15 crc kubenswrapper[4967]: I1008 00:07:15.459592 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:15Z","lastTransitionTime":"2025-10-08T00:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:15 crc kubenswrapper[4967]: I1008 00:07:15.469858 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58c4704-a88a-43b4-b944-4de626f13f7e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22f5dabb39da0a7bd01db4dae1437b9a231a1fcc0dfe749a4a5a637a72cf33c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01a2b3dc785520c79eae998482e5a1347725c76584357073238fe60a619edb77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6aa6f5e81feceab3d1c7587152a70546b365e629a69267e4d3084324a1de764a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63c2b7c4b2791bd1f16888c58b878dbff1f87884e68c9949de914d0d3e1d39fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:15Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:15 crc kubenswrapper[4967]: I1008 00:07:15.484721 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:15Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:15 crc kubenswrapper[4967]: I1008 00:07:15.505502 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9mrr7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9299bc4f-f8d3-4baf-9885-39573982fd64\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://994beafece9dead0d85ae74dc453a142460705fd66b52ba22f2fa799cf27da08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb49c8476b482416588c4d403016940730064a619c0641eb029c947a4bab1215\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb49c8476b482416588c4d403016940730064a619c0641eb029c947a4bab1215\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b7b63962dcbf477f1fa4d71826a4aa3358920a2c1b2f5e48716411b5bdca049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b7b63962dcbf477f1fa4d71826a4aa3358920a2c1b2f5e48716411b5bdca049\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1e770dc737173fd4f3220ed6c52b422ccbd829722386a37029e1c3e10d80e1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1e770dc737173fd4f3220ed6c52b422ccbd829722386a37029e1c3e10d80e1a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e357d970878c996f219a79f1876b363176c9847b9575f33090f62bc299529a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e357d970878c996f219a79f1876b363176c9847b9575f33090f62bc299529a95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f098966bfad49d06278eaa72be042596d488115e597737d9a0a0cfc78f99f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49f098966bfad49d06278eaa72be042596d488115e597737d9a0a0cfc78f99f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1d53876b7b863c612d597fd716c3260e8857b6542802073bb730f6756071b46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c1d53876b7b863c612d597fd716c3260e8857b6542802073bb730f6756071b46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9mrr7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:15Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:15 crc kubenswrapper[4967]: I1008 00:07:15.522775 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fa4b747-ab6e-4a11-9903-ef2dcb9668c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbadcdade274b6d9be940f22bc949e3c55cadf95dcad43c8cb7a1085c0e6708c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d0a0b2ef4a98ee828688081a74f4b7ecc31c8fae415d7257a560ecd43960ce3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96b53b1b283025b3d078a3f505d469d1872ad4bb8ab8e612c36c70b0b00a920\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1400e110997c64fa33d6b8672c7e91cab5f782952ddf8e1dea2a882689026e7f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1400e110997c64fa33d6b8672c7e91cab5f782952ddf8e1dea2a882689026e7f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 00:06:56.760088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 00:06:56.761210 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-369805205/tls.crt::/tmp/serving-cert-369805205/tls.key\\\\\\\"\\\\nI1008 00:07:02.321448 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 00:07:02.325932 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 00:07:02.325980 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 00:07:02.326025 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 00:07:02.326042 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 00:07:02.336203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 00:07:02.336240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 00:07:02.336245 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 00:07:02.336250 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 00:07:02.336253 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 00:07:02.336256 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 00:07:02.336258 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 00:07:02.336430 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 00:07:02.340774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:46Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca28d8db19036e47a7ef3dadb70941f6c6cf5735e7703977d036185c46fce40\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5b222ad84c31d8746dfe03124c7287b401deb3ca2c62ce2b4fbd40e638cee0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5b222ad84c31d8746dfe03124c7287b401deb3ca2c62ce2b4fbd40e638cee0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:15Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:15 crc kubenswrapper[4967]: I1008 00:07:15.538563 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:15Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:15 crc kubenswrapper[4967]: I1008 00:07:15.553961 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbfcdc882e2be1cb9cf666c9e7775671078f7dbbfe8421dd15db737b89920a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:15Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:15 crc kubenswrapper[4967]: I1008 00:07:15.562351 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:15 crc kubenswrapper[4967]: I1008 00:07:15.562392 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:15 crc kubenswrapper[4967]: I1008 00:07:15.562402 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:15 crc kubenswrapper[4967]: I1008 00:07:15.562419 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:15 crc kubenswrapper[4967]: I1008 00:07:15.562430 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:15Z","lastTransitionTime":"2025-10-08T00:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:15 crc kubenswrapper[4967]: I1008 00:07:15.570491 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9461162a-c27d-465e-9ca7-2d3936877cd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9778d78ddc7325abd4d77bc5e71b0b114f50be5dda4a438c7cb6b05640c8a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z9tp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28760af04ef321d0f91b2bce3ad074d44b63dabf676cc37f3be6aa8726341a2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z9tp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xvdtp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:15Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:15 crc kubenswrapper[4967]: I1008 00:07:15.584171 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94e63bc5641ce80d507f76a19b203f78c42368a7a4105f3fd09e19af21a1b735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:15Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:15 crc kubenswrapper[4967]: I1008 00:07:15.599096 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e98666fd3152bab57d3723e7da72aae2703b79f7ca73eb3c3e5f0e062a82592\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cf1cbe5548146fe359aafa351864125ec6b8bdd7b796aa4d311f6f3a16ad212\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:15Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:15 crc kubenswrapper[4967]: I1008 00:07:15.611225 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9tdbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41264a12-9486-4fee-a08c-607b590e5a70\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://806e85df796ae300d279f73a0cc4433a5bb1f7620ee8ce33c8dc4674c4ce0596\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5blv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9tdbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:15Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:15 crc kubenswrapper[4967]: I1008 00:07:15.642130 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f1af2ba-8907-4d14-afa1-ffde207c80f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4535cffb8c9d64e662e282ecb44f6817b4604866f0bb75015f3b888d5554f821\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b03b83ddff429dbfa8fe052901aaa86a39ff0dd75477ffa8a6b6f4198f5819ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c9fa0559eb35ba67e11e36069ac8d0b2fec76399788eef668a08ad4353df3ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3312263e12bd42fbed0476ab5ace66b5497f62c9477865e130d67e3922e05cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e69dc603055aa221d44d11c049ddcc50cec568979911a310ed773a90341a499\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115f7a33f3dedae8f59d2f45f465f95f48b99c824e895ef7a0be9fe7a1b61d23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9972551ade22d86d9ccd63097a793a2fb4ce74779a99f4b9aab1cf215087975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ead9cad55ea37f9a14c0beddb89065098e821e6b43db7e5ddacd89a2a12bf36a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T00:07:13Z\\\",\\\"message\\\":\\\"tory.go:140\\\\nI1008 00:07:13.601660 6243 reflector.go:311] Stopping reflector *v1.ClusterUserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1008 00:07:13.601605 6243 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1008 00:07:13.601819 6243 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1008 00:07:13.601862 6243 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1008 00:07:13.601897 6243 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1008 00:07:13.601932 6243 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1008 00:07:13.601973 6243 factory.go:656] Stopping watch factory\\\\nI1008 00:07:13.602003 6243 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1008 00:07:13.601622 6243 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1008 00:07:13.602415 6243 handler.go:208] Removed *v1.Node event handler 7\\\\nI1008 00:07:13.602437 6243 handler.go:208] Removed *v1.Node event handler 2\\\\nI1008 00:07:13.603125 6243 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1008 00:07:13.603186 6243 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1008 00:07:13.603406 6243 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:10Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27fb2b1225e386dcd7fecaaa0aa54ba1fdfafabfde6c3824ae07315313067542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bs8w9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:15Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:15 crc kubenswrapper[4967]: I1008 00:07:15.664194 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-v5n7t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abe7ae2d-7e63-4249-96e1-4000e9628ccc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52575e853ff6727dde7c89e4add4362be126127db27191c2a7bf2871efa719d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpsjd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:06Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-v5n7t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:15Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:15 crc kubenswrapper[4967]: I1008 00:07:15.666454 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:15 crc kubenswrapper[4967]: I1008 00:07:15.666515 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:15 crc kubenswrapper[4967]: I1008 00:07:15.666536 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:15 crc kubenswrapper[4967]: I1008 00:07:15.666564 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:15 crc kubenswrapper[4967]: I1008 00:07:15.666587 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:15Z","lastTransitionTime":"2025-10-08T00:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:15 crc kubenswrapper[4967]: I1008 00:07:15.680224 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:15Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:15 crc kubenswrapper[4967]: I1008 00:07:15.694966 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-npz2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c71501b1-aa9e-420e-b1bd-7e15365e7adc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dfe6c420d804b2d7daef4b5cfe3f87b77b369287c4d65553fa93f5f24f589d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gblk2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-npz2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:15Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:15 crc kubenswrapper[4967]: I1008 00:07:15.769770 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:15 crc kubenswrapper[4967]: I1008 00:07:15.769840 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:15 crc kubenswrapper[4967]: I1008 00:07:15.769859 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:15 crc kubenswrapper[4967]: I1008 00:07:15.769881 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:15 crc kubenswrapper[4967]: I1008 00:07:15.769897 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:15Z","lastTransitionTime":"2025-10-08T00:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:15 crc kubenswrapper[4967]: I1008 00:07:15.872982 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:15 crc kubenswrapper[4967]: I1008 00:07:15.873044 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:15 crc kubenswrapper[4967]: I1008 00:07:15.873063 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:15 crc kubenswrapper[4967]: I1008 00:07:15.873086 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:15 crc kubenswrapper[4967]: I1008 00:07:15.873104 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:15Z","lastTransitionTime":"2025-10-08T00:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:15 crc kubenswrapper[4967]: I1008 00:07:15.975872 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:15 crc kubenswrapper[4967]: I1008 00:07:15.975928 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:15 crc kubenswrapper[4967]: I1008 00:07:15.975946 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:15 crc kubenswrapper[4967]: I1008 00:07:15.975971 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:15 crc kubenswrapper[4967]: I1008 00:07:15.975987 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:15Z","lastTransitionTime":"2025-10-08T00:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:16 crc kubenswrapper[4967]: I1008 00:07:16.079497 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:16 crc kubenswrapper[4967]: I1008 00:07:16.079558 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:16 crc kubenswrapper[4967]: I1008 00:07:16.079578 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:16 crc kubenswrapper[4967]: I1008 00:07:16.079605 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:16 crc kubenswrapper[4967]: I1008 00:07:16.079622 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:16Z","lastTransitionTime":"2025-10-08T00:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:16 crc kubenswrapper[4967]: I1008 00:07:16.117444 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 00:07:16 crc kubenswrapper[4967]: E1008 00:07:16.117637 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 00:07:16 crc kubenswrapper[4967]: I1008 00:07:16.183404 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:16 crc kubenswrapper[4967]: I1008 00:07:16.183476 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:16 crc kubenswrapper[4967]: I1008 00:07:16.183501 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:16 crc kubenswrapper[4967]: I1008 00:07:16.183532 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:16 crc kubenswrapper[4967]: I1008 00:07:16.183552 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:16Z","lastTransitionTime":"2025-10-08T00:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:16 crc kubenswrapper[4967]: I1008 00:07:16.287649 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:16 crc kubenswrapper[4967]: I1008 00:07:16.287868 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:16 crc kubenswrapper[4967]: I1008 00:07:16.287929 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:16 crc kubenswrapper[4967]: I1008 00:07:16.287961 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:16 crc kubenswrapper[4967]: I1008 00:07:16.287984 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:16Z","lastTransitionTime":"2025-10-08T00:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:16 crc kubenswrapper[4967]: I1008 00:07:16.391396 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:16 crc kubenswrapper[4967]: I1008 00:07:16.391454 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:16 crc kubenswrapper[4967]: I1008 00:07:16.391506 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:16 crc kubenswrapper[4967]: I1008 00:07:16.391536 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:16 crc kubenswrapper[4967]: I1008 00:07:16.391558 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:16Z","lastTransitionTime":"2025-10-08T00:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:16 crc kubenswrapper[4967]: I1008 00:07:16.454796 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bs8w9_4f1af2ba-8907-4d14-afa1-ffde207c80f4/ovnkube-controller/1.log" Oct 08 00:07:16 crc kubenswrapper[4967]: I1008 00:07:16.455822 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bs8w9_4f1af2ba-8907-4d14-afa1-ffde207c80f4/ovnkube-controller/0.log" Oct 08 00:07:16 crc kubenswrapper[4967]: I1008 00:07:16.463576 4967 generic.go:334] "Generic (PLEG): container finished" podID="4f1af2ba-8907-4d14-afa1-ffde207c80f4" containerID="f9972551ade22d86d9ccd63097a793a2fb4ce74779a99f4b9aab1cf215087975" exitCode=1 Oct 08 00:07:16 crc kubenswrapper[4967]: I1008 00:07:16.463633 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" event={"ID":"4f1af2ba-8907-4d14-afa1-ffde207c80f4","Type":"ContainerDied","Data":"f9972551ade22d86d9ccd63097a793a2fb4ce74779a99f4b9aab1cf215087975"} Oct 08 00:07:16 crc kubenswrapper[4967]: I1008 00:07:16.463696 4967 scope.go:117] "RemoveContainer" containerID="ead9cad55ea37f9a14c0beddb89065098e821e6b43db7e5ddacd89a2a12bf36a" Oct 08 00:07:16 crc kubenswrapper[4967]: I1008 00:07:16.464979 4967 scope.go:117] "RemoveContainer" containerID="f9972551ade22d86d9ccd63097a793a2fb4ce74779a99f4b9aab1cf215087975" Oct 08 00:07:16 crc kubenswrapper[4967]: E1008 00:07:16.465343 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-bs8w9_openshift-ovn-kubernetes(4f1af2ba-8907-4d14-afa1-ffde207c80f4)\"" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" podUID="4f1af2ba-8907-4d14-afa1-ffde207c80f4" Oct 08 00:07:16 crc kubenswrapper[4967]: I1008 00:07:16.489477 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-npz2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c71501b1-aa9e-420e-b1bd-7e15365e7adc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dfe6c420d804b2d7daef4b5cfe3f87b77b369287c4d65553fa93f5f24f589d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gblk2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-npz2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:16Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:16 crc kubenswrapper[4967]: I1008 00:07:16.494401 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:16 crc kubenswrapper[4967]: I1008 00:07:16.494440 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:16 crc kubenswrapper[4967]: I1008 00:07:16.494455 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:16 crc kubenswrapper[4967]: I1008 00:07:16.494473 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:16 crc kubenswrapper[4967]: I1008 00:07:16.494487 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:16Z","lastTransitionTime":"2025-10-08T00:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:16 crc kubenswrapper[4967]: I1008 00:07:16.510765 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:16Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:16 crc kubenswrapper[4967]: I1008 00:07:16.528091 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:16Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:16 crc kubenswrapper[4967]: I1008 00:07:16.548270 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9mrr7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9299bc4f-f8d3-4baf-9885-39573982fd64\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://994beafece9dead0d85ae74dc453a142460705fd66b52ba22f2fa799cf27da08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb49c8476b482416588c4d403016940730064a619c0641eb029c947a4bab1215\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb49c8476b482416588c4d403016940730064a619c0641eb029c947a4bab1215\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b7b63962dcbf477f1fa4d71826a4aa3358920a2c1b2f5e48716411b5bdca049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b7b63962dcbf477f1fa4d71826a4aa3358920a2c1b2f5e48716411b5bdca049\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1e770dc737173fd4f3220ed6c52b422ccbd829722386a37029e1c3e10d80e1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1e770dc737173fd4f3220ed6c52b422ccbd829722386a37029e1c3e10d80e1a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e357d970878c996f219a79f1876b363176c9847b9575f33090f62bc299529a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e357d970878c996f219a79f1876b363176c9847b9575f33090f62bc299529a95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f098966bfad49d06278eaa72be042596d488115e597737d9a0a0cfc78f99f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49f098966bfad49d06278eaa72be042596d488115e597737d9a0a0cfc78f99f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1d53876b7b863c612d597fd716c3260e8857b6542802073bb730f6756071b46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c1d53876b7b863c612d597fd716c3260e8857b6542802073bb730f6756071b46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9mrr7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:16Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:16 crc kubenswrapper[4967]: I1008 00:07:16.564113 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58c4704-a88a-43b4-b944-4de626f13f7e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22f5dabb39da0a7bd01db4dae1437b9a231a1fcc0dfe749a4a5a637a72cf33c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01a2b3dc785520c79eae998482e5a1347725c76584357073238fe60a619edb77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6aa6f5e81feceab3d1c7587152a70546b365e629a69267e4d3084324a1de764a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63c2b7c4b2791bd1f16888c58b878dbff1f87884e68c9949de914d0d3e1d39fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:16Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:16 crc kubenswrapper[4967]: I1008 00:07:16.584105 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:16Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:16 crc kubenswrapper[4967]: I1008 00:07:16.597197 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:16 crc kubenswrapper[4967]: I1008 00:07:16.597231 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:16 crc kubenswrapper[4967]: I1008 00:07:16.597243 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:16 crc kubenswrapper[4967]: I1008 00:07:16.597259 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:16 crc kubenswrapper[4967]: I1008 00:07:16.597273 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:16Z","lastTransitionTime":"2025-10-08T00:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:16 crc kubenswrapper[4967]: I1008 00:07:16.608502 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbfcdc882e2be1cb9cf666c9e7775671078f7dbbfe8421dd15db737b89920a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:16Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:16 crc kubenswrapper[4967]: I1008 00:07:16.630870 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9461162a-c27d-465e-9ca7-2d3936877cd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9778d78ddc7325abd4d77bc5e71b0b114f50be5dda4a438c7cb6b05640c8a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z9tp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28760af04ef321d0f91b2bce3ad074d44b63dabf676cc37f3be6aa8726341a2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z9tp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xvdtp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:16Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:16 crc kubenswrapper[4967]: I1008 00:07:16.645589 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fa4b747-ab6e-4a11-9903-ef2dcb9668c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbadcdade274b6d9be940f22bc949e3c55cadf95dcad43c8cb7a1085c0e6708c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d0a0b2ef4a98ee828688081a74f4b7ecc31c8fae415d7257a560ecd43960ce3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96b53b1b283025b3d078a3f505d469d1872ad4bb8ab8e612c36c70b0b00a920\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1400e110997c64fa33d6b8672c7e91cab5f782952ddf8e1dea2a882689026e7f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1400e110997c64fa33d6b8672c7e91cab5f782952ddf8e1dea2a882689026e7f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 00:06:56.760088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 00:06:56.761210 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-369805205/tls.crt::/tmp/serving-cert-369805205/tls.key\\\\\\\"\\\\nI1008 00:07:02.321448 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 00:07:02.325932 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 00:07:02.325980 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 00:07:02.326025 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 00:07:02.326042 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 00:07:02.336203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 00:07:02.336240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 00:07:02.336245 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 00:07:02.336250 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 00:07:02.336253 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 00:07:02.336256 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 00:07:02.336258 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 00:07:02.336430 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 00:07:02.340774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:46Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca28d8db19036e47a7ef3dadb70941f6c6cf5735e7703977d036185c46fce40\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5b222ad84c31d8746dfe03124c7287b401deb3ca2c62ce2b4fbd40e638cee0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5b222ad84c31d8746dfe03124c7287b401deb3ca2c62ce2b4fbd40e638cee0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:16Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:16 crc kubenswrapper[4967]: I1008 00:07:16.657928 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9tdbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41264a12-9486-4fee-a08c-607b590e5a70\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://806e85df796ae300d279f73a0cc4433a5bb1f7620ee8ce33c8dc4674c4ce0596\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5blv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9tdbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:16Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:16 crc kubenswrapper[4967]: I1008 00:07:16.683626 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f1af2ba-8907-4d14-afa1-ffde207c80f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4535cffb8c9d64e662e282ecb44f6817b4604866f0bb75015f3b888d5554f821\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b03b83ddff429dbfa8fe052901aaa86a39ff0dd75477ffa8a6b6f4198f5819ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c9fa0559eb35ba67e11e36069ac8d0b2fec76399788eef668a08ad4353df3ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3312263e12bd42fbed0476ab5ace66b5497f62c9477865e130d67e3922e05cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e69dc603055aa221d44d11c049ddcc50cec568979911a310ed773a90341a499\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115f7a33f3dedae8f59d2f45f465f95f48b99c824e895ef7a0be9fe7a1b61d23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9972551ade22d86d9ccd63097a793a2fb4ce74779a99f4b9aab1cf215087975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ead9cad55ea37f9a14c0beddb89065098e821e6b43db7e5ddacd89a2a12bf36a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T00:07:13Z\\\",\\\"message\\\":\\\"tory.go:140\\\\nI1008 00:07:13.601660 6243 reflector.go:311] Stopping reflector *v1.ClusterUserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1008 00:07:13.601605 6243 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1008 00:07:13.601819 6243 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1008 00:07:13.601862 6243 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1008 00:07:13.601897 6243 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1008 00:07:13.601932 6243 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1008 00:07:13.601973 6243 factory.go:656] Stopping watch factory\\\\nI1008 00:07:13.602003 6243 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1008 00:07:13.601622 6243 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1008 00:07:13.602415 6243 handler.go:208] Removed *v1.Node event handler 7\\\\nI1008 00:07:13.602437 6243 handler.go:208] Removed *v1.Node event handler 2\\\\nI1008 00:07:13.603125 6243 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1008 00:07:13.603186 6243 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1008 00:07:13.603406 6243 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:10Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9972551ade22d86d9ccd63097a793a2fb4ce74779a99f4b9aab1cf215087975\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T00:07:15Z\\\",\\\"message\\\":\\\"ring admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:15Z is after 2025-08-24T17:21:41Z]\\\\nI1008 00:07:15.446035 6404 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-marketplace/marketplace-operator-metrics]} name:Service_openshift-marketplace/marketplace-operator-metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.53:8081: 10.217.5.53:8383:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {89fe421e-04e8-4967-ac75-77a0e6f784ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1008 00:07:15.446063 6404 obj_retry.go:303] Retry object s\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27fb2b1225e386dcd7fecaaa0aa54ba1fdfafabfde6c3824ae07315313067542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bs8w9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:16Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:16 crc kubenswrapper[4967]: I1008 00:07:16.725766 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-v5n7t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abe7ae2d-7e63-4249-96e1-4000e9628ccc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52575e853ff6727dde7c89e4add4362be126127db27191c2a7bf2871efa719d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpsjd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:06Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-v5n7t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:16Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:16 crc kubenswrapper[4967]: I1008 00:07:16.727265 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:16 crc kubenswrapper[4967]: I1008 00:07:16.727307 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:16 crc kubenswrapper[4967]: I1008 00:07:16.727318 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:16 crc kubenswrapper[4967]: I1008 00:07:16.727333 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:16 crc kubenswrapper[4967]: I1008 00:07:16.727345 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:16Z","lastTransitionTime":"2025-10-08T00:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:16 crc kubenswrapper[4967]: I1008 00:07:16.744532 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94e63bc5641ce80d507f76a19b203f78c42368a7a4105f3fd09e19af21a1b735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:16Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:16 crc kubenswrapper[4967]: I1008 00:07:16.760599 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e98666fd3152bab57d3723e7da72aae2703b79f7ca73eb3c3e5f0e062a82592\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cf1cbe5548146fe359aafa351864125ec6b8bdd7b796aa4d311f6f3a16ad212\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:16Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:16 crc kubenswrapper[4967]: I1008 00:07:16.829486 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:16 crc kubenswrapper[4967]: I1008 00:07:16.829533 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:16 crc kubenswrapper[4967]: I1008 00:07:16.829545 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:16 crc kubenswrapper[4967]: I1008 00:07:16.829564 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:16 crc kubenswrapper[4967]: I1008 00:07:16.829578 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:16Z","lastTransitionTime":"2025-10-08T00:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:16 crc kubenswrapper[4967]: I1008 00:07:16.932759 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:16 crc kubenswrapper[4967]: I1008 00:07:16.932826 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:16 crc kubenswrapper[4967]: I1008 00:07:16.932852 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:16 crc kubenswrapper[4967]: I1008 00:07:16.932882 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:16 crc kubenswrapper[4967]: I1008 00:07:16.932903 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:16Z","lastTransitionTime":"2025-10-08T00:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.013353 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kpmlg"] Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.013878 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kpmlg" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.018826 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.019825 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.039799 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.040104 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.040227 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.040344 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.040465 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:17Z","lastTransitionTime":"2025-10-08T00:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.043045 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58c4704-a88a-43b4-b944-4de626f13f7e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22f5dabb39da0a7bd01db4dae1437b9a231a1fcc0dfe749a4a5a637a72cf33c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01a2b3dc785520c79eae998482e5a1347725c76584357073238fe60a619edb77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6aa6f5e81feceab3d1c7587152a70546b365e629a69267e4d3084324a1de764a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63c2b7c4b2791bd1f16888c58b878dbff1f87884e68c9949de914d0d3e1d39fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:17Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.059958 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:17Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.079212 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9mrr7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9299bc4f-f8d3-4baf-9885-39573982fd64\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://994beafece9dead0d85ae74dc453a142460705fd66b52ba22f2fa799cf27da08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb49c8476b482416588c4d403016940730064a619c0641eb029c947a4bab1215\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb49c8476b482416588c4d403016940730064a619c0641eb029c947a4bab1215\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b7b63962dcbf477f1fa4d71826a4aa3358920a2c1b2f5e48716411b5bdca049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b7b63962dcbf477f1fa4d71826a4aa3358920a2c1b2f5e48716411b5bdca049\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1e770dc737173fd4f3220ed6c52b422ccbd829722386a37029e1c3e10d80e1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1e770dc737173fd4f3220ed6c52b422ccbd829722386a37029e1c3e10d80e1a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e357d970878c996f219a79f1876b363176c9847b9575f33090f62bc299529a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e357d970878c996f219a79f1876b363176c9847b9575f33090f62bc299529a95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f098966bfad49d06278eaa72be042596d488115e597737d9a0a0cfc78f99f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49f098966bfad49d06278eaa72be042596d488115e597737d9a0a0cfc78f99f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1d53876b7b863c612d597fd716c3260e8857b6542802073bb730f6756071b46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c1d53876b7b863c612d597fd716c3260e8857b6542802073bb730f6756071b46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9mrr7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:17Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.092694 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kpmlg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f975d47d-42a6-422e-80e2-d1ee282817fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qznqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qznqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-kpmlg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:17Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.107276 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fa4b747-ab6e-4a11-9903-ef2dcb9668c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbadcdade274b6d9be940f22bc949e3c55cadf95dcad43c8cb7a1085c0e6708c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d0a0b2ef4a98ee828688081a74f4b7ecc31c8fae415d7257a560ecd43960ce3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96b53b1b283025b3d078a3f505d469d1872ad4bb8ab8e612c36c70b0b00a920\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1400e110997c64fa33d6b8672c7e91cab5f782952ddf8e1dea2a882689026e7f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1400e110997c64fa33d6b8672c7e91cab5f782952ddf8e1dea2a882689026e7f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 00:06:56.760088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 00:06:56.761210 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-369805205/tls.crt::/tmp/serving-cert-369805205/tls.key\\\\\\\"\\\\nI1008 00:07:02.321448 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 00:07:02.325932 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 00:07:02.325980 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 00:07:02.326025 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 00:07:02.326042 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 00:07:02.336203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 00:07:02.336240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 00:07:02.336245 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 00:07:02.336250 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 00:07:02.336253 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 00:07:02.336256 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 00:07:02.336258 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 00:07:02.336430 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 00:07:02.340774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:46Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca28d8db19036e47a7ef3dadb70941f6c6cf5735e7703977d036185c46fce40\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5b222ad84c31d8746dfe03124c7287b401deb3ca2c62ce2b4fbd40e638cee0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5b222ad84c31d8746dfe03124c7287b401deb3ca2c62ce2b4fbd40e638cee0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:17Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.117304 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.117385 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 00:07:17 crc kubenswrapper[4967]: E1008 00:07:17.117446 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 00:07:17 crc kubenswrapper[4967]: E1008 00:07:17.117518 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.118314 4967 scope.go:117] "RemoveContainer" containerID="1400e110997c64fa33d6b8672c7e91cab5f782952ddf8e1dea2a882689026e7f" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.120599 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:17Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.137378 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f975d47d-42a6-422e-80e2-d1ee282817fd-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-kpmlg\" (UID: \"f975d47d-42a6-422e-80e2-d1ee282817fd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kpmlg" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.137454 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f975d47d-42a6-422e-80e2-d1ee282817fd-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-kpmlg\" (UID: \"f975d47d-42a6-422e-80e2-d1ee282817fd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kpmlg" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.137487 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qznqw\" (UniqueName: \"kubernetes.io/projected/f975d47d-42a6-422e-80e2-d1ee282817fd-kube-api-access-qznqw\") pod \"ovnkube-control-plane-749d76644c-kpmlg\" (UID: \"f975d47d-42a6-422e-80e2-d1ee282817fd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kpmlg" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.137558 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f975d47d-42a6-422e-80e2-d1ee282817fd-env-overrides\") pod \"ovnkube-control-plane-749d76644c-kpmlg\" (UID: \"f975d47d-42a6-422e-80e2-d1ee282817fd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kpmlg" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.138422 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbfcdc882e2be1cb9cf666c9e7775671078f7dbbfe8421dd15db737b89920a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:17Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.143834 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.143906 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.143931 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.143966 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.143984 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:17Z","lastTransitionTime":"2025-10-08T00:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.155889 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9461162a-c27d-465e-9ca7-2d3936877cd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9778d78ddc7325abd4d77bc5e71b0b114f50be5dda4a438c7cb6b05640c8a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z9tp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28760af04ef321d0f91b2bce3ad074d44b63dabf676cc37f3be6aa8726341a2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z9tp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xvdtp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:17Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.171635 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94e63bc5641ce80d507f76a19b203f78c42368a7a4105f3fd09e19af21a1b735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:17Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.186508 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e98666fd3152bab57d3723e7da72aae2703b79f7ca73eb3c3e5f0e062a82592\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cf1cbe5548146fe359aafa351864125ec6b8bdd7b796aa4d311f6f3a16ad212\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:17Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.200484 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9tdbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41264a12-9486-4fee-a08c-607b590e5a70\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://806e85df796ae300d279f73a0cc4433a5bb1f7620ee8ce33c8dc4674c4ce0596\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5blv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9tdbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:17Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.221704 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f1af2ba-8907-4d14-afa1-ffde207c80f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4535cffb8c9d64e662e282ecb44f6817b4604866f0bb75015f3b888d5554f821\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b03b83ddff429dbfa8fe052901aaa86a39ff0dd75477ffa8a6b6f4198f5819ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c9fa0559eb35ba67e11e36069ac8d0b2fec76399788eef668a08ad4353df3ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3312263e12bd42fbed0476ab5ace66b5497f62c9477865e130d67e3922e05cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e69dc603055aa221d44d11c049ddcc50cec568979911a310ed773a90341a499\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115f7a33f3dedae8f59d2f45f465f95f48b99c824e895ef7a0be9fe7a1b61d23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9972551ade22d86d9ccd63097a793a2fb4ce74779a99f4b9aab1cf215087975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ead9cad55ea37f9a14c0beddb89065098e821e6b43db7e5ddacd89a2a12bf36a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T00:07:13Z\\\",\\\"message\\\":\\\"tory.go:140\\\\nI1008 00:07:13.601660 6243 reflector.go:311] Stopping reflector *v1.ClusterUserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1008 00:07:13.601605 6243 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1008 00:07:13.601819 6243 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1008 00:07:13.601862 6243 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1008 00:07:13.601897 6243 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1008 00:07:13.601932 6243 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1008 00:07:13.601973 6243 factory.go:656] Stopping watch factory\\\\nI1008 00:07:13.602003 6243 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1008 00:07:13.601622 6243 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1008 00:07:13.602415 6243 handler.go:208] Removed *v1.Node event handler 7\\\\nI1008 00:07:13.602437 6243 handler.go:208] Removed *v1.Node event handler 2\\\\nI1008 00:07:13.603125 6243 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1008 00:07:13.603186 6243 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1008 00:07:13.603406 6243 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:10Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9972551ade22d86d9ccd63097a793a2fb4ce74779a99f4b9aab1cf215087975\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T00:07:15Z\\\",\\\"message\\\":\\\"ring admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:15Z is after 2025-08-24T17:21:41Z]\\\\nI1008 00:07:15.446035 6404 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-marketplace/marketplace-operator-metrics]} name:Service_openshift-marketplace/marketplace-operator-metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.53:8081: 10.217.5.53:8383:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {89fe421e-04e8-4967-ac75-77a0e6f784ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1008 00:07:15.446063 6404 obj_retry.go:303] Retry object s\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27fb2b1225e386dcd7fecaaa0aa54ba1fdfafabfde6c3824ae07315313067542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bs8w9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:17Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.233199 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-v5n7t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abe7ae2d-7e63-4249-96e1-4000e9628ccc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52575e853ff6727dde7c89e4add4362be126127db27191c2a7bf2871efa719d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpsjd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:06Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-v5n7t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:17Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.238771 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f975d47d-42a6-422e-80e2-d1ee282817fd-env-overrides\") pod \"ovnkube-control-plane-749d76644c-kpmlg\" (UID: \"f975d47d-42a6-422e-80e2-d1ee282817fd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kpmlg" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.238945 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f975d47d-42a6-422e-80e2-d1ee282817fd-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-kpmlg\" (UID: \"f975d47d-42a6-422e-80e2-d1ee282817fd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kpmlg" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.239062 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f975d47d-42a6-422e-80e2-d1ee282817fd-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-kpmlg\" (UID: \"f975d47d-42a6-422e-80e2-d1ee282817fd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kpmlg" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.239169 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qznqw\" (UniqueName: \"kubernetes.io/projected/f975d47d-42a6-422e-80e2-d1ee282817fd-kube-api-access-qznqw\") pod \"ovnkube-control-plane-749d76644c-kpmlg\" (UID: \"f975d47d-42a6-422e-80e2-d1ee282817fd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kpmlg" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.240118 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f975d47d-42a6-422e-80e2-d1ee282817fd-env-overrides\") pod \"ovnkube-control-plane-749d76644c-kpmlg\" (UID: \"f975d47d-42a6-422e-80e2-d1ee282817fd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kpmlg" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.240130 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f975d47d-42a6-422e-80e2-d1ee282817fd-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-kpmlg\" (UID: \"f975d47d-42a6-422e-80e2-d1ee282817fd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kpmlg" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.245402 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:17Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.246410 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f975d47d-42a6-422e-80e2-d1ee282817fd-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-kpmlg\" (UID: \"f975d47d-42a6-422e-80e2-d1ee282817fd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kpmlg" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.247182 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.247207 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.247216 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.247231 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.247240 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:17Z","lastTransitionTime":"2025-10-08T00:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.260007 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-npz2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c71501b1-aa9e-420e-b1bd-7e15365e7adc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dfe6c420d804b2d7daef4b5cfe3f87b77b369287c4d65553fa93f5f24f589d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gblk2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-npz2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:17Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.265844 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qznqw\" (UniqueName: \"kubernetes.io/projected/f975d47d-42a6-422e-80e2-d1ee282817fd-kube-api-access-qznqw\") pod \"ovnkube-control-plane-749d76644c-kpmlg\" (UID: \"f975d47d-42a6-422e-80e2-d1ee282817fd\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kpmlg" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.332033 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kpmlg" Oct 08 00:07:17 crc kubenswrapper[4967]: W1008 00:07:17.347765 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf975d47d_42a6_422e_80e2_d1ee282817fd.slice/crio-a348a7e04cbb7146c06d0f7aa4cf08bc4cf9637c4a52738c5fbe0316c4830ad6 WatchSource:0}: Error finding container a348a7e04cbb7146c06d0f7aa4cf08bc4cf9637c4a52738c5fbe0316c4830ad6: Status 404 returned error can't find the container with id a348a7e04cbb7146c06d0f7aa4cf08bc4cf9637c4a52738c5fbe0316c4830ad6 Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.349740 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.349798 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.349818 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.349843 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.349861 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:17Z","lastTransitionTime":"2025-10-08T00:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.451940 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.451984 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.451997 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.452016 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.452028 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:17Z","lastTransitionTime":"2025-10-08T00:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.468422 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bs8w9_4f1af2ba-8907-4d14-afa1-ffde207c80f4/ovnkube-controller/1.log" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.472876 4967 scope.go:117] "RemoveContainer" containerID="f9972551ade22d86d9ccd63097a793a2fb4ce74779a99f4b9aab1cf215087975" Oct 08 00:07:17 crc kubenswrapper[4967]: E1008 00:07:17.473164 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-bs8w9_openshift-ovn-kubernetes(4f1af2ba-8907-4d14-afa1-ffde207c80f4)\"" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" podUID="4f1af2ba-8907-4d14-afa1-ffde207c80f4" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.476920 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.479693 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"ef86a2422616e1e0d5c3afe844818c2bdf2150f941d1aefc506995f58587b2bd"} Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.480251 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.481188 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kpmlg" event={"ID":"f975d47d-42a6-422e-80e2-d1ee282817fd","Type":"ContainerStarted","Data":"a348a7e04cbb7146c06d0f7aa4cf08bc4cf9637c4a52738c5fbe0316c4830ad6"} Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.492543 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-npz2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c71501b1-aa9e-420e-b1bd-7e15365e7adc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dfe6c420d804b2d7daef4b5cfe3f87b77b369287c4d65553fa93f5f24f589d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gblk2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-npz2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:17Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.511164 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:17Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.525057 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:17Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.541426 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9mrr7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9299bc4f-f8d3-4baf-9885-39573982fd64\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://994beafece9dead0d85ae74dc453a142460705fd66b52ba22f2fa799cf27da08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb49c8476b482416588c4d403016940730064a619c0641eb029c947a4bab1215\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb49c8476b482416588c4d403016940730064a619c0641eb029c947a4bab1215\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b7b63962dcbf477f1fa4d71826a4aa3358920a2c1b2f5e48716411b5bdca049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b7b63962dcbf477f1fa4d71826a4aa3358920a2c1b2f5e48716411b5bdca049\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1e770dc737173fd4f3220ed6c52b422ccbd829722386a37029e1c3e10d80e1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1e770dc737173fd4f3220ed6c52b422ccbd829722386a37029e1c3e10d80e1a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e357d970878c996f219a79f1876b363176c9847b9575f33090f62bc299529a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e357d970878c996f219a79f1876b363176c9847b9575f33090f62bc299529a95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f098966bfad49d06278eaa72be042596d488115e597737d9a0a0cfc78f99f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49f098966bfad49d06278eaa72be042596d488115e597737d9a0a0cfc78f99f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1d53876b7b863c612d597fd716c3260e8857b6542802073bb730f6756071b46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c1d53876b7b863c612d597fd716c3260e8857b6542802073bb730f6756071b46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9mrr7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:17Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.553061 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kpmlg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f975d47d-42a6-422e-80e2-d1ee282817fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qznqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qznqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-kpmlg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:17Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.556088 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.556189 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.556209 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.556237 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.556256 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:17Z","lastTransitionTime":"2025-10-08T00:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.565787 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58c4704-a88a-43b4-b944-4de626f13f7e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22f5dabb39da0a7bd01db4dae1437b9a231a1fcc0dfe749a4a5a637a72cf33c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01a2b3dc785520c79eae998482e5a1347725c76584357073238fe60a619edb77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6aa6f5e81feceab3d1c7587152a70546b365e629a69267e4d3084324a1de764a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63c2b7c4b2791bd1f16888c58b878dbff1f87884e68c9949de914d0d3e1d39fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:17Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.578924 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:17Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.595044 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbfcdc882e2be1cb9cf666c9e7775671078f7dbbfe8421dd15db737b89920a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:17Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.610409 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9461162a-c27d-465e-9ca7-2d3936877cd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9778d78ddc7325abd4d77bc5e71b0b114f50be5dda4a438c7cb6b05640c8a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z9tp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28760af04ef321d0f91b2bce3ad074d44b63dabf676cc37f3be6aa8726341a2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z9tp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xvdtp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:17Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.632449 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fa4b747-ab6e-4a11-9903-ef2dcb9668c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbadcdade274b6d9be940f22bc949e3c55cadf95dcad43c8cb7a1085c0e6708c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d0a0b2ef4a98ee828688081a74f4b7ecc31c8fae415d7257a560ecd43960ce3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96b53b1b283025b3d078a3f505d469d1872ad4bb8ab8e612c36c70b0b00a920\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1400e110997c64fa33d6b8672c7e91cab5f782952ddf8e1dea2a882689026e7f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1400e110997c64fa33d6b8672c7e91cab5f782952ddf8e1dea2a882689026e7f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 00:06:56.760088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 00:06:56.761210 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-369805205/tls.crt::/tmp/serving-cert-369805205/tls.key\\\\\\\"\\\\nI1008 00:07:02.321448 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 00:07:02.325932 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 00:07:02.325980 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 00:07:02.326025 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 00:07:02.326042 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 00:07:02.336203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 00:07:02.336240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 00:07:02.336245 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 00:07:02.336250 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 00:07:02.336253 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 00:07:02.336256 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 00:07:02.336258 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 00:07:02.336430 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 00:07:02.340774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:46Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca28d8db19036e47a7ef3dadb70941f6c6cf5735e7703977d036185c46fce40\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5b222ad84c31d8746dfe03124c7287b401deb3ca2c62ce2b4fbd40e638cee0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5b222ad84c31d8746dfe03124c7287b401deb3ca2c62ce2b4fbd40e638cee0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:17Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.647062 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9tdbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41264a12-9486-4fee-a08c-607b590e5a70\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://806e85df796ae300d279f73a0cc4433a5bb1f7620ee8ce33c8dc4674c4ce0596\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5blv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9tdbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:17Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.660407 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.660457 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.660485 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.660515 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.660533 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:17Z","lastTransitionTime":"2025-10-08T00:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.679191 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f1af2ba-8907-4d14-afa1-ffde207c80f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4535cffb8c9d64e662e282ecb44f6817b4604866f0bb75015f3b888d5554f821\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b03b83ddff429dbfa8fe052901aaa86a39ff0dd75477ffa8a6b6f4198f5819ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c9fa0559eb35ba67e11e36069ac8d0b2fec76399788eef668a08ad4353df3ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3312263e12bd42fbed0476ab5ace66b5497f62c9477865e130d67e3922e05cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e69dc603055aa221d44d11c049ddcc50cec568979911a310ed773a90341a499\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115f7a33f3dedae8f59d2f45f465f95f48b99c824e895ef7a0be9fe7a1b61d23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9972551ade22d86d9ccd63097a793a2fb4ce74779a99f4b9aab1cf215087975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9972551ade22d86d9ccd63097a793a2fb4ce74779a99f4b9aab1cf215087975\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T00:07:15Z\\\",\\\"message\\\":\\\"ring admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:15Z is after 2025-08-24T17:21:41Z]\\\\nI1008 00:07:15.446035 6404 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-marketplace/marketplace-operator-metrics]} name:Service_openshift-marketplace/marketplace-operator-metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.53:8081: 10.217.5.53:8383:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {89fe421e-04e8-4967-ac75-77a0e6f784ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1008 00:07:15.446063 6404 obj_retry.go:303] Retry object s\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:14Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-bs8w9_openshift-ovn-kubernetes(4f1af2ba-8907-4d14-afa1-ffde207c80f4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27fb2b1225e386dcd7fecaaa0aa54ba1fdfafabfde6c3824ae07315313067542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bs8w9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:17Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.694039 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-v5n7t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abe7ae2d-7e63-4249-96e1-4000e9628ccc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52575e853ff6727dde7c89e4add4362be126127db27191c2a7bf2871efa719d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpsjd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:06Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-v5n7t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:17Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.718034 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94e63bc5641ce80d507f76a19b203f78c42368a7a4105f3fd09e19af21a1b735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:17Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.736776 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e98666fd3152bab57d3723e7da72aae2703b79f7ca73eb3c3e5f0e062a82592\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cf1cbe5548146fe359aafa351864125ec6b8bdd7b796aa4d311f6f3a16ad212\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:17Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.748821 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-v5n7t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abe7ae2d-7e63-4249-96e1-4000e9628ccc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52575e853ff6727dde7c89e4add4362be126127db27191c2a7bf2871efa719d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpsjd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:06Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-v5n7t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:17Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.764413 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.764481 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.764502 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.764532 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.764552 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:17Z","lastTransitionTime":"2025-10-08T00:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.776424 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94e63bc5641ce80d507f76a19b203f78c42368a7a4105f3fd09e19af21a1b735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:17Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.792774 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e98666fd3152bab57d3723e7da72aae2703b79f7ca73eb3c3e5f0e062a82592\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cf1cbe5548146fe359aafa351864125ec6b8bdd7b796aa4d311f6f3a16ad212\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:17Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.808080 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9tdbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41264a12-9486-4fee-a08c-607b590e5a70\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://806e85df796ae300d279f73a0cc4433a5bb1f7620ee8ce33c8dc4674c4ce0596\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5blv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9tdbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:17Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.832615 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f1af2ba-8907-4d14-afa1-ffde207c80f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4535cffb8c9d64e662e282ecb44f6817b4604866f0bb75015f3b888d5554f821\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b03b83ddff429dbfa8fe052901aaa86a39ff0dd75477ffa8a6b6f4198f5819ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c9fa0559eb35ba67e11e36069ac8d0b2fec76399788eef668a08ad4353df3ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3312263e12bd42fbed0476ab5ace66b5497f62c9477865e130d67e3922e05cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e69dc603055aa221d44d11c049ddcc50cec568979911a310ed773a90341a499\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115f7a33f3dedae8f59d2f45f465f95f48b99c824e895ef7a0be9fe7a1b61d23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9972551ade22d86d9ccd63097a793a2fb4ce74779a99f4b9aab1cf215087975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9972551ade22d86d9ccd63097a793a2fb4ce74779a99f4b9aab1cf215087975\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T00:07:15Z\\\",\\\"message\\\":\\\"ring admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:15Z is after 2025-08-24T17:21:41Z]\\\\nI1008 00:07:15.446035 6404 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-marketplace/marketplace-operator-metrics]} name:Service_openshift-marketplace/marketplace-operator-metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.53:8081: 10.217.5.53:8383:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {89fe421e-04e8-4967-ac75-77a0e6f784ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1008 00:07:15.446063 6404 obj_retry.go:303] Retry object s\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:14Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-bs8w9_openshift-ovn-kubernetes(4f1af2ba-8907-4d14-afa1-ffde207c80f4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27fb2b1225e386dcd7fecaaa0aa54ba1fdfafabfde6c3824ae07315313067542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bs8w9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:17Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.844937 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:17Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.858244 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-npz2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c71501b1-aa9e-420e-b1bd-7e15365e7adc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dfe6c420d804b2d7daef4b5cfe3f87b77b369287c4d65553fa93f5f24f589d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gblk2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-npz2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:17Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.867036 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.867084 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.867102 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.867126 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.867143 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:17Z","lastTransitionTime":"2025-10-08T00:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.883100 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kpmlg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f975d47d-42a6-422e-80e2-d1ee282817fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qznqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qznqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-kpmlg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:17Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.900379 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58c4704-a88a-43b4-b944-4de626f13f7e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22f5dabb39da0a7bd01db4dae1437b9a231a1fcc0dfe749a4a5a637a72cf33c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01a2b3dc785520c79eae998482e5a1347725c76584357073238fe60a619edb77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6aa6f5e81feceab3d1c7587152a70546b365e629a69267e4d3084324a1de764a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63c2b7c4b2791bd1f16888c58b878dbff1f87884e68c9949de914d0d3e1d39fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:17Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.916064 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:17Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.936468 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9mrr7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9299bc4f-f8d3-4baf-9885-39573982fd64\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://994beafece9dead0d85ae74dc453a142460705fd66b52ba22f2fa799cf27da08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb49c8476b482416588c4d403016940730064a619c0641eb029c947a4bab1215\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb49c8476b482416588c4d403016940730064a619c0641eb029c947a4bab1215\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b7b63962dcbf477f1fa4d71826a4aa3358920a2c1b2f5e48716411b5bdca049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b7b63962dcbf477f1fa4d71826a4aa3358920a2c1b2f5e48716411b5bdca049\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1e770dc737173fd4f3220ed6c52b422ccbd829722386a37029e1c3e10d80e1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1e770dc737173fd4f3220ed6c52b422ccbd829722386a37029e1c3e10d80e1a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e357d970878c996f219a79f1876b363176c9847b9575f33090f62bc299529a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e357d970878c996f219a79f1876b363176c9847b9575f33090f62bc299529a95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f098966bfad49d06278eaa72be042596d488115e597737d9a0a0cfc78f99f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49f098966bfad49d06278eaa72be042596d488115e597737d9a0a0cfc78f99f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1d53876b7b863c612d597fd716c3260e8857b6542802073bb730f6756071b46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c1d53876b7b863c612d597fd716c3260e8857b6542802073bb730f6756071b46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9mrr7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:17Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.948158 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9461162a-c27d-465e-9ca7-2d3936877cd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9778d78ddc7325abd4d77bc5e71b0b114f50be5dda4a438c7cb6b05640c8a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z9tp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28760af04ef321d0f91b2bce3ad074d44b63dabf676cc37f3be6aa8726341a2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z9tp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xvdtp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:17Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.967575 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fa4b747-ab6e-4a11-9903-ef2dcb9668c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbadcdade274b6d9be940f22bc949e3c55cadf95dcad43c8cb7a1085c0e6708c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d0a0b2ef4a98ee828688081a74f4b7ecc31c8fae415d7257a560ecd43960ce3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96b53b1b283025b3d078a3f505d469d1872ad4bb8ab8e612c36c70b0b00a920\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef86a2422616e1e0d5c3afe844818c2bdf2150f941d1aefc506995f58587b2bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1400e110997c64fa33d6b8672c7e91cab5f782952ddf8e1dea2a882689026e7f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 00:06:56.760088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 00:06:56.761210 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-369805205/tls.crt::/tmp/serving-cert-369805205/tls.key\\\\\\\"\\\\nI1008 00:07:02.321448 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 00:07:02.325932 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 00:07:02.325980 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 00:07:02.326025 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 00:07:02.326042 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 00:07:02.336203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 00:07:02.336240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 00:07:02.336245 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 00:07:02.336250 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 00:07:02.336253 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 00:07:02.336256 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 00:07:02.336258 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 00:07:02.336430 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 00:07:02.340774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:46Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca28d8db19036e47a7ef3dadb70941f6c6cf5735e7703977d036185c46fce40\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5b222ad84c31d8746dfe03124c7287b401deb3ca2c62ce2b4fbd40e638cee0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5b222ad84c31d8746dfe03124c7287b401deb3ca2c62ce2b4fbd40e638cee0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:17Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.969842 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.969865 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.969874 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.969889 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.969898 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:17Z","lastTransitionTime":"2025-10-08T00:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.980513 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:17Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:17 crc kubenswrapper[4967]: I1008 00:07:17.990973 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbfcdc882e2be1cb9cf666c9e7775671078f7dbbfe8421dd15db737b89920a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:17Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:18 crc kubenswrapper[4967]: I1008 00:07:18.071911 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:18 crc kubenswrapper[4967]: I1008 00:07:18.072178 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:18 crc kubenswrapper[4967]: I1008 00:07:18.072266 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:18 crc kubenswrapper[4967]: I1008 00:07:18.072393 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:18 crc kubenswrapper[4967]: I1008 00:07:18.072477 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:18Z","lastTransitionTime":"2025-10-08T00:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:18 crc kubenswrapper[4967]: I1008 00:07:18.117345 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 00:07:18 crc kubenswrapper[4967]: E1008 00:07:18.117469 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 00:07:18 crc kubenswrapper[4967]: I1008 00:07:18.175651 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:18 crc kubenswrapper[4967]: I1008 00:07:18.175706 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:18 crc kubenswrapper[4967]: I1008 00:07:18.175717 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:18 crc kubenswrapper[4967]: I1008 00:07:18.175735 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:18 crc kubenswrapper[4967]: I1008 00:07:18.175747 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:18Z","lastTransitionTime":"2025-10-08T00:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:18 crc kubenswrapper[4967]: I1008 00:07:18.277520 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:18 crc kubenswrapper[4967]: I1008 00:07:18.277558 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:18 crc kubenswrapper[4967]: I1008 00:07:18.277570 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:18 crc kubenswrapper[4967]: I1008 00:07:18.277585 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:18 crc kubenswrapper[4967]: I1008 00:07:18.277595 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:18Z","lastTransitionTime":"2025-10-08T00:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:18 crc kubenswrapper[4967]: I1008 00:07:18.379957 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:18 crc kubenswrapper[4967]: I1008 00:07:18.380730 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:18 crc kubenswrapper[4967]: I1008 00:07:18.380839 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:18 crc kubenswrapper[4967]: I1008 00:07:18.380927 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:18 crc kubenswrapper[4967]: I1008 00:07:18.381028 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:18Z","lastTransitionTime":"2025-10-08T00:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:18 crc kubenswrapper[4967]: I1008 00:07:18.483726 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:18 crc kubenswrapper[4967]: I1008 00:07:18.484042 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:18 crc kubenswrapper[4967]: I1008 00:07:18.484175 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:18 crc kubenswrapper[4967]: I1008 00:07:18.484282 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:18 crc kubenswrapper[4967]: I1008 00:07:18.484355 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:18Z","lastTransitionTime":"2025-10-08T00:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:18 crc kubenswrapper[4967]: I1008 00:07:18.487129 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kpmlg" event={"ID":"f975d47d-42a6-422e-80e2-d1ee282817fd","Type":"ContainerStarted","Data":"7ef61854aea40924c5372b0553cd0b7e8faa39a1d42fec5b65cf867c368f72c5"} Oct 08 00:07:18 crc kubenswrapper[4967]: I1008 00:07:18.487271 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kpmlg" event={"ID":"f975d47d-42a6-422e-80e2-d1ee282817fd","Type":"ContainerStarted","Data":"a317ae6640eb38b3aee0f9738cca890e8ab3d92075b620cd312f40a0628d46c1"} Oct 08 00:07:18 crc kubenswrapper[4967]: I1008 00:07:18.508068 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94e63bc5641ce80d507f76a19b203f78c42368a7a4105f3fd09e19af21a1b735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:18Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:18 crc kubenswrapper[4967]: I1008 00:07:18.522302 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e98666fd3152bab57d3723e7da72aae2703b79f7ca73eb3c3e5f0e062a82592\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cf1cbe5548146fe359aafa351864125ec6b8bdd7b796aa4d311f6f3a16ad212\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:18Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:18 crc kubenswrapper[4967]: I1008 00:07:18.534001 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9tdbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41264a12-9486-4fee-a08c-607b590e5a70\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://806e85df796ae300d279f73a0cc4433a5bb1f7620ee8ce33c8dc4674c4ce0596\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5blv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9tdbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:18Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:18 crc kubenswrapper[4967]: I1008 00:07:18.573056 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f1af2ba-8907-4d14-afa1-ffde207c80f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4535cffb8c9d64e662e282ecb44f6817b4604866f0bb75015f3b888d5554f821\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b03b83ddff429dbfa8fe052901aaa86a39ff0dd75477ffa8a6b6f4198f5819ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c9fa0559eb35ba67e11e36069ac8d0b2fec76399788eef668a08ad4353df3ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3312263e12bd42fbed0476ab5ace66b5497f62c9477865e130d67e3922e05cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e69dc603055aa221d44d11c049ddcc50cec568979911a310ed773a90341a499\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115f7a33f3dedae8f59d2f45f465f95f48b99c824e895ef7a0be9fe7a1b61d23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9972551ade22d86d9ccd63097a793a2fb4ce74779a99f4b9aab1cf215087975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9972551ade22d86d9ccd63097a793a2fb4ce74779a99f4b9aab1cf215087975\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T00:07:15Z\\\",\\\"message\\\":\\\"ring admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:15Z is after 2025-08-24T17:21:41Z]\\\\nI1008 00:07:15.446035 6404 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-marketplace/marketplace-operator-metrics]} name:Service_openshift-marketplace/marketplace-operator-metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.53:8081: 10.217.5.53:8383:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {89fe421e-04e8-4967-ac75-77a0e6f784ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1008 00:07:15.446063 6404 obj_retry.go:303] Retry object s\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:14Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-bs8w9_openshift-ovn-kubernetes(4f1af2ba-8907-4d14-afa1-ffde207c80f4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27fb2b1225e386dcd7fecaaa0aa54ba1fdfafabfde6c3824ae07315313067542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bs8w9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:18Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:18 crc kubenswrapper[4967]: I1008 00:07:18.587036 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:18 crc kubenswrapper[4967]: I1008 00:07:18.587097 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:18 crc kubenswrapper[4967]: I1008 00:07:18.587116 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:18 crc kubenswrapper[4967]: I1008 00:07:18.587139 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:18 crc kubenswrapper[4967]: I1008 00:07:18.587155 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:18Z","lastTransitionTime":"2025-10-08T00:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:18 crc kubenswrapper[4967]: I1008 00:07:18.588914 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-v5n7t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abe7ae2d-7e63-4249-96e1-4000e9628ccc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52575e853ff6727dde7c89e4add4362be126127db27191c2a7bf2871efa719d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpsjd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:06Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-v5n7t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:18Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:18 crc kubenswrapper[4967]: I1008 00:07:18.607124 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:18Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:18 crc kubenswrapper[4967]: I1008 00:07:18.627199 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-npz2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c71501b1-aa9e-420e-b1bd-7e15365e7adc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dfe6c420d804b2d7daef4b5cfe3f87b77b369287c4d65553fa93f5f24f589d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gblk2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-npz2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:18Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:18 crc kubenswrapper[4967]: I1008 00:07:18.654438 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58c4704-a88a-43b4-b944-4de626f13f7e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22f5dabb39da0a7bd01db4dae1437b9a231a1fcc0dfe749a4a5a637a72cf33c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01a2b3dc785520c79eae998482e5a1347725c76584357073238fe60a619edb77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6aa6f5e81feceab3d1c7587152a70546b365e629a69267e4d3084324a1de764a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63c2b7c4b2791bd1f16888c58b878dbff1f87884e68c9949de914d0d3e1d39fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:18Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:18 crc kubenswrapper[4967]: I1008 00:07:18.683641 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:18Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:18 crc kubenswrapper[4967]: I1008 00:07:18.689223 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:18 crc kubenswrapper[4967]: I1008 00:07:18.689262 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:18 crc kubenswrapper[4967]: I1008 00:07:18.689273 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:18 crc kubenswrapper[4967]: I1008 00:07:18.689289 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:18 crc kubenswrapper[4967]: I1008 00:07:18.689301 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:18Z","lastTransitionTime":"2025-10-08T00:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:18 crc kubenswrapper[4967]: I1008 00:07:18.712128 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9mrr7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9299bc4f-f8d3-4baf-9885-39573982fd64\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://994beafece9dead0d85ae74dc453a142460705fd66b52ba22f2fa799cf27da08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb49c8476b482416588c4d403016940730064a619c0641eb029c947a4bab1215\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb49c8476b482416588c4d403016940730064a619c0641eb029c947a4bab1215\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b7b63962dcbf477f1fa4d71826a4aa3358920a2c1b2f5e48716411b5bdca049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b7b63962dcbf477f1fa4d71826a4aa3358920a2c1b2f5e48716411b5bdca049\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1e770dc737173fd4f3220ed6c52b422ccbd829722386a37029e1c3e10d80e1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1e770dc737173fd4f3220ed6c52b422ccbd829722386a37029e1c3e10d80e1a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e357d970878c996f219a79f1876b363176c9847b9575f33090f62bc299529a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e357d970878c996f219a79f1876b363176c9847b9575f33090f62bc299529a95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f098966bfad49d06278eaa72be042596d488115e597737d9a0a0cfc78f99f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49f098966bfad49d06278eaa72be042596d488115e597737d9a0a0cfc78f99f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1d53876b7b863c612d597fd716c3260e8857b6542802073bb730f6756071b46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c1d53876b7b863c612d597fd716c3260e8857b6542802073bb730f6756071b46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9mrr7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:18Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:18 crc kubenswrapper[4967]: I1008 00:07:18.728919 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kpmlg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f975d47d-42a6-422e-80e2-d1ee282817fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a317ae6640eb38b3aee0f9738cca890e8ab3d92075b620cd312f40a0628d46c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qznqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ef61854aea40924c5372b0553cd0b7e8faa39a1d42fec5b65cf867c368f72c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qznqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-kpmlg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:18Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:18 crc kubenswrapper[4967]: I1008 00:07:18.742981 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fa4b747-ab6e-4a11-9903-ef2dcb9668c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbadcdade274b6d9be940f22bc949e3c55cadf95dcad43c8cb7a1085c0e6708c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d0a0b2ef4a98ee828688081a74f4b7ecc31c8fae415d7257a560ecd43960ce3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96b53b1b283025b3d078a3f505d469d1872ad4bb8ab8e612c36c70b0b00a920\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef86a2422616e1e0d5c3afe844818c2bdf2150f941d1aefc506995f58587b2bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1400e110997c64fa33d6b8672c7e91cab5f782952ddf8e1dea2a882689026e7f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 00:06:56.760088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 00:06:56.761210 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-369805205/tls.crt::/tmp/serving-cert-369805205/tls.key\\\\\\\"\\\\nI1008 00:07:02.321448 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 00:07:02.325932 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 00:07:02.325980 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 00:07:02.326025 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 00:07:02.326042 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 00:07:02.336203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 00:07:02.336240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 00:07:02.336245 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 00:07:02.336250 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 00:07:02.336253 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 00:07:02.336256 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 00:07:02.336258 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 00:07:02.336430 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 00:07:02.340774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:46Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca28d8db19036e47a7ef3dadb70941f6c6cf5735e7703977d036185c46fce40\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5b222ad84c31d8746dfe03124c7287b401deb3ca2c62ce2b4fbd40e638cee0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5b222ad84c31d8746dfe03124c7287b401deb3ca2c62ce2b4fbd40e638cee0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:18Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:18 crc kubenswrapper[4967]: I1008 00:07:18.759202 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:18Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:18 crc kubenswrapper[4967]: I1008 00:07:18.770800 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbfcdc882e2be1cb9cf666c9e7775671078f7dbbfe8421dd15db737b89920a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:18Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:18 crc kubenswrapper[4967]: I1008 00:07:18.784284 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9461162a-c27d-465e-9ca7-2d3936877cd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9778d78ddc7325abd4d77bc5e71b0b114f50be5dda4a438c7cb6b05640c8a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z9tp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28760af04ef321d0f91b2bce3ad074d44b63dabf676cc37f3be6aa8726341a2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z9tp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xvdtp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:18Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:18 crc kubenswrapper[4967]: I1008 00:07:18.792129 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:18 crc kubenswrapper[4967]: I1008 00:07:18.792175 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:18 crc kubenswrapper[4967]: I1008 00:07:18.792185 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:18 crc kubenswrapper[4967]: I1008 00:07:18.792201 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:18 crc kubenswrapper[4967]: I1008 00:07:18.792211 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:18Z","lastTransitionTime":"2025-10-08T00:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:18 crc kubenswrapper[4967]: I1008 00:07:18.894864 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:18 crc kubenswrapper[4967]: I1008 00:07:18.894923 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:18 crc kubenswrapper[4967]: I1008 00:07:18.894943 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:18 crc kubenswrapper[4967]: I1008 00:07:18.894968 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:18 crc kubenswrapper[4967]: I1008 00:07:18.894988 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:18Z","lastTransitionTime":"2025-10-08T00:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:18 crc kubenswrapper[4967]: I1008 00:07:18.927855 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-2h7z8"] Oct 08 00:07:18 crc kubenswrapper[4967]: I1008 00:07:18.928731 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2h7z8" Oct 08 00:07:18 crc kubenswrapper[4967]: E1008 00:07:18.928859 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2h7z8" podUID="59f68771-e25d-4367-87ee-68a31d767882" Oct 08 00:07:18 crc kubenswrapper[4967]: I1008 00:07:18.948330 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:18Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:18 crc kubenswrapper[4967]: I1008 00:07:18.958186 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 00:07:18 crc kubenswrapper[4967]: I1008 00:07:18.958290 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 00:07:18 crc kubenswrapper[4967]: I1008 00:07:18.958327 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 00:07:18 crc kubenswrapper[4967]: I1008 00:07:18.958372 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 00:07:18 crc kubenswrapper[4967]: E1008 00:07:18.958413 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 00:07:34.958385294 +0000 UTC m=+53.608919963 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:07:18 crc kubenswrapper[4967]: E1008 00:07:18.958466 4967 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 00:07:18 crc kubenswrapper[4967]: I1008 00:07:18.958476 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 00:07:18 crc kubenswrapper[4967]: E1008 00:07:18.958492 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 00:07:18 crc kubenswrapper[4967]: E1008 00:07:18.958521 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 00:07:34.958504948 +0000 UTC m=+53.609039697 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 00:07:18 crc kubenswrapper[4967]: E1008 00:07:18.958528 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 00:07:18 crc kubenswrapper[4967]: E1008 00:07:18.958542 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 00:07:18 crc kubenswrapper[4967]: E1008 00:07:18.958580 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 00:07:18 crc kubenswrapper[4967]: E1008 00:07:18.958604 4967 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 00:07:18 crc kubenswrapper[4967]: E1008 00:07:18.958631 4967 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 00:07:18 crc kubenswrapper[4967]: E1008 00:07:18.958550 4967 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 00:07:18 crc kubenswrapper[4967]: E1008 00:07:18.958663 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-08 00:07:34.958644251 +0000 UTC m=+53.609178940 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 00:07:18 crc kubenswrapper[4967]: E1008 00:07:18.958730 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 00:07:34.958717393 +0000 UTC m=+53.609252072 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 00:07:18 crc kubenswrapper[4967]: E1008 00:07:18.958755 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-08 00:07:34.958743204 +0000 UTC m=+53.609277893 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 00:07:18 crc kubenswrapper[4967]: I1008 00:07:18.969425 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:18 crc kubenswrapper[4967]: I1008 00:07:18.969475 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:18 crc kubenswrapper[4967]: I1008 00:07:18.969487 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:18 crc kubenswrapper[4967]: I1008 00:07:18.969507 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:18 crc kubenswrapper[4967]: I1008 00:07:18.969523 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:18Z","lastTransitionTime":"2025-10-08T00:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:18 crc kubenswrapper[4967]: I1008 00:07:18.975938 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-npz2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c71501b1-aa9e-420e-b1bd-7e15365e7adc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dfe6c420d804b2d7daef4b5cfe3f87b77b369287c4d65553fa93f5f24f589d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gblk2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-npz2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:18Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:18 crc kubenswrapper[4967]: E1008 00:07:18.990573 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5cddad4c-22d1-42d7-a373-87b47c1cc97b\\\",\\\"systemUUID\\\":\\\"b243ca1c-20cd-4e6f-96b2-52fce74dd2d4\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:18Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:18 crc kubenswrapper[4967]: I1008 00:07:18.994013 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2h7z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59f68771-e25d-4367-87ee-68a31d767882\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7r2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7r2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:18Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2h7z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:18Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:18 crc kubenswrapper[4967]: I1008 00:07:18.996044 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:18 crc kubenswrapper[4967]: I1008 00:07:18.996086 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:18 crc kubenswrapper[4967]: I1008 00:07:18.996098 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:18 crc kubenswrapper[4967]: I1008 00:07:18.996116 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:18 crc kubenswrapper[4967]: I1008 00:07:18.996131 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:18Z","lastTransitionTime":"2025-10-08T00:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:19 crc kubenswrapper[4967]: I1008 00:07:19.009474 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58c4704-a88a-43b4-b944-4de626f13f7e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22f5dabb39da0a7bd01db4dae1437b9a231a1fcc0dfe749a4a5a637a72cf33c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01a2b3dc785520c79eae998482e5a1347725c76584357073238fe60a619edb77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6aa6f5e81feceab3d1c7587152a70546b365e629a69267e4d3084324a1de764a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63c2b7c4b2791bd1f16888c58b878dbff1f87884e68c9949de914d0d3e1d39fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:19Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:19 crc kubenswrapper[4967]: E1008 00:07:19.013198 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5cddad4c-22d1-42d7-a373-87b47c1cc97b\\\",\\\"systemUUID\\\":\\\"b243ca1c-20cd-4e6f-96b2-52fce74dd2d4\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:19Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:19 crc kubenswrapper[4967]: I1008 00:07:19.017800 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:19 crc kubenswrapper[4967]: I1008 00:07:19.017887 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:19 crc kubenswrapper[4967]: I1008 00:07:19.017909 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:19 crc kubenswrapper[4967]: I1008 00:07:19.017936 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:19 crc kubenswrapper[4967]: I1008 00:07:19.017955 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:19Z","lastTransitionTime":"2025-10-08T00:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:19 crc kubenswrapper[4967]: I1008 00:07:19.026415 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:19Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:19 crc kubenswrapper[4967]: E1008 00:07:19.038059 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5cddad4c-22d1-42d7-a373-87b47c1cc97b\\\",\\\"systemUUID\\\":\\\"b243ca1c-20cd-4e6f-96b2-52fce74dd2d4\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:19Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:19 crc kubenswrapper[4967]: I1008 00:07:19.042505 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:19 crc kubenswrapper[4967]: I1008 00:07:19.042572 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:19 crc kubenswrapper[4967]: I1008 00:07:19.042597 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:19 crc kubenswrapper[4967]: I1008 00:07:19.042624 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:19 crc kubenswrapper[4967]: I1008 00:07:19.042646 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:19Z","lastTransitionTime":"2025-10-08T00:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:19 crc kubenswrapper[4967]: I1008 00:07:19.054482 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9mrr7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9299bc4f-f8d3-4baf-9885-39573982fd64\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://994beafece9dead0d85ae74dc453a142460705fd66b52ba22f2fa799cf27da08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb49c8476b482416588c4d403016940730064a619c0641eb029c947a4bab1215\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb49c8476b482416588c4d403016940730064a619c0641eb029c947a4bab1215\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b7b63962dcbf477f1fa4d71826a4aa3358920a2c1b2f5e48716411b5bdca049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b7b63962dcbf477f1fa4d71826a4aa3358920a2c1b2f5e48716411b5bdca049\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1e770dc737173fd4f3220ed6c52b422ccbd829722386a37029e1c3e10d80e1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1e770dc737173fd4f3220ed6c52b422ccbd829722386a37029e1c3e10d80e1a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e357d970878c996f219a79f1876b363176c9847b9575f33090f62bc299529a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e357d970878c996f219a79f1876b363176c9847b9575f33090f62bc299529a95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f098966bfad49d06278eaa72be042596d488115e597737d9a0a0cfc78f99f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49f098966bfad49d06278eaa72be042596d488115e597737d9a0a0cfc78f99f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1d53876b7b863c612d597fd716c3260e8857b6542802073bb730f6756071b46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c1d53876b7b863c612d597fd716c3260e8857b6542802073bb730f6756071b46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9mrr7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:19Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:19 crc kubenswrapper[4967]: E1008 00:07:19.057111 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5cddad4c-22d1-42d7-a373-87b47c1cc97b\\\",\\\"systemUUID\\\":\\\"b243ca1c-20cd-4e6f-96b2-52fce74dd2d4\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:19Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:19 crc kubenswrapper[4967]: I1008 00:07:19.059481 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/59f68771-e25d-4367-87ee-68a31d767882-metrics-certs\") pod \"network-metrics-daemon-2h7z8\" (UID: \"59f68771-e25d-4367-87ee-68a31d767882\") " pod="openshift-multus/network-metrics-daemon-2h7z8" Oct 08 00:07:19 crc kubenswrapper[4967]: I1008 00:07:19.059534 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c7r2h\" (UniqueName: \"kubernetes.io/projected/59f68771-e25d-4367-87ee-68a31d767882-kube-api-access-c7r2h\") pod \"network-metrics-daemon-2h7z8\" (UID: \"59f68771-e25d-4367-87ee-68a31d767882\") " pod="openshift-multus/network-metrics-daemon-2h7z8" Oct 08 00:07:19 crc kubenswrapper[4967]: I1008 00:07:19.062523 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:19 crc kubenswrapper[4967]: I1008 00:07:19.062568 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:19 crc kubenswrapper[4967]: I1008 00:07:19.062585 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:19 crc kubenswrapper[4967]: I1008 00:07:19.062607 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:19 crc kubenswrapper[4967]: I1008 00:07:19.062623 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:19Z","lastTransitionTime":"2025-10-08T00:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:19 crc kubenswrapper[4967]: I1008 00:07:19.073906 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kpmlg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f975d47d-42a6-422e-80e2-d1ee282817fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a317ae6640eb38b3aee0f9738cca890e8ab3d92075b620cd312f40a0628d46c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qznqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ef61854aea40924c5372b0553cd0b7e8faa39a1d42fec5b65cf867c368f72c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qznqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-kpmlg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:19Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:19 crc kubenswrapper[4967]: E1008 00:07:19.081433 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5cddad4c-22d1-42d7-a373-87b47c1cc97b\\\",\\\"systemUUID\\\":\\\"b243ca1c-20cd-4e6f-96b2-52fce74dd2d4\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:19Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:19 crc kubenswrapper[4967]: E1008 00:07:19.081722 4967 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 08 00:07:19 crc kubenswrapper[4967]: I1008 00:07:19.083973 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:19 crc kubenswrapper[4967]: I1008 00:07:19.084024 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:19 crc kubenswrapper[4967]: I1008 00:07:19.084039 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:19 crc kubenswrapper[4967]: I1008 00:07:19.084063 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:19 crc kubenswrapper[4967]: I1008 00:07:19.084077 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:19Z","lastTransitionTime":"2025-10-08T00:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:19 crc kubenswrapper[4967]: I1008 00:07:19.096232 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fa4b747-ab6e-4a11-9903-ef2dcb9668c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbadcdade274b6d9be940f22bc949e3c55cadf95dcad43c8cb7a1085c0e6708c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d0a0b2ef4a98ee828688081a74f4b7ecc31c8fae415d7257a560ecd43960ce3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96b53b1b283025b3d078a3f505d469d1872ad4bb8ab8e612c36c70b0b00a920\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef86a2422616e1e0d5c3afe844818c2bdf2150f941d1aefc506995f58587b2bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1400e110997c64fa33d6b8672c7e91cab5f782952ddf8e1dea2a882689026e7f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 00:06:56.760088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 00:06:56.761210 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-369805205/tls.crt::/tmp/serving-cert-369805205/tls.key\\\\\\\"\\\\nI1008 00:07:02.321448 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 00:07:02.325932 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 00:07:02.325980 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 00:07:02.326025 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 00:07:02.326042 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 00:07:02.336203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 00:07:02.336240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 00:07:02.336245 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 00:07:02.336250 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 00:07:02.336253 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 00:07:02.336256 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 00:07:02.336258 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 00:07:02.336430 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 00:07:02.340774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:46Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca28d8db19036e47a7ef3dadb70941f6c6cf5735e7703977d036185c46fce40\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5b222ad84c31d8746dfe03124c7287b401deb3ca2c62ce2b4fbd40e638cee0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5b222ad84c31d8746dfe03124c7287b401deb3ca2c62ce2b4fbd40e638cee0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:19Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:19 crc kubenswrapper[4967]: I1008 00:07:19.113597 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:19Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:19 crc kubenswrapper[4967]: I1008 00:07:19.116896 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 00:07:19 crc kubenswrapper[4967]: I1008 00:07:19.116911 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 00:07:19 crc kubenswrapper[4967]: E1008 00:07:19.117088 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 00:07:19 crc kubenswrapper[4967]: E1008 00:07:19.117006 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 00:07:19 crc kubenswrapper[4967]: I1008 00:07:19.131105 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbfcdc882e2be1cb9cf666c9e7775671078f7dbbfe8421dd15db737b89920a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:19Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:19 crc kubenswrapper[4967]: I1008 00:07:19.145976 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9461162a-c27d-465e-9ca7-2d3936877cd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9778d78ddc7325abd4d77bc5e71b0b114f50be5dda4a438c7cb6b05640c8a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z9tp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28760af04ef321d0f91b2bce3ad074d44b63dabf676cc37f3be6aa8726341a2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z9tp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xvdtp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:19Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:19 crc kubenswrapper[4967]: I1008 00:07:19.161215 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/59f68771-e25d-4367-87ee-68a31d767882-metrics-certs\") pod \"network-metrics-daemon-2h7z8\" (UID: \"59f68771-e25d-4367-87ee-68a31d767882\") " pod="openshift-multus/network-metrics-daemon-2h7z8" Oct 08 00:07:19 crc kubenswrapper[4967]: I1008 00:07:19.161303 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c7r2h\" (UniqueName: \"kubernetes.io/projected/59f68771-e25d-4367-87ee-68a31d767882-kube-api-access-c7r2h\") pod \"network-metrics-daemon-2h7z8\" (UID: \"59f68771-e25d-4367-87ee-68a31d767882\") " pod="openshift-multus/network-metrics-daemon-2h7z8" Oct 08 00:07:19 crc kubenswrapper[4967]: E1008 00:07:19.161394 4967 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 00:07:19 crc kubenswrapper[4967]: E1008 00:07:19.161467 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/59f68771-e25d-4367-87ee-68a31d767882-metrics-certs podName:59f68771-e25d-4367-87ee-68a31d767882 nodeName:}" failed. No retries permitted until 2025-10-08 00:07:19.661449663 +0000 UTC m=+38.311984322 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/59f68771-e25d-4367-87ee-68a31d767882-metrics-certs") pod "network-metrics-daemon-2h7z8" (UID: "59f68771-e25d-4367-87ee-68a31d767882") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 00:07:19 crc kubenswrapper[4967]: I1008 00:07:19.164959 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94e63bc5641ce80d507f76a19b203f78c42368a7a4105f3fd09e19af21a1b735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:19Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:19 crc kubenswrapper[4967]: I1008 00:07:19.186871 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:19 crc kubenswrapper[4967]: I1008 00:07:19.187101 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:19 crc kubenswrapper[4967]: I1008 00:07:19.187115 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:19 crc kubenswrapper[4967]: I1008 00:07:19.187131 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:19 crc kubenswrapper[4967]: I1008 00:07:19.187142 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:19Z","lastTransitionTime":"2025-10-08T00:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:19 crc kubenswrapper[4967]: I1008 00:07:19.187984 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e98666fd3152bab57d3723e7da72aae2703b79f7ca73eb3c3e5f0e062a82592\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cf1cbe5548146fe359aafa351864125ec6b8bdd7b796aa4d311f6f3a16ad212\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:19Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:19 crc kubenswrapper[4967]: I1008 00:07:19.191615 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c7r2h\" (UniqueName: \"kubernetes.io/projected/59f68771-e25d-4367-87ee-68a31d767882-kube-api-access-c7r2h\") pod \"network-metrics-daemon-2h7z8\" (UID: \"59f68771-e25d-4367-87ee-68a31d767882\") " pod="openshift-multus/network-metrics-daemon-2h7z8" Oct 08 00:07:19 crc kubenswrapper[4967]: I1008 00:07:19.206291 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9tdbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41264a12-9486-4fee-a08c-607b590e5a70\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://806e85df796ae300d279f73a0cc4433a5bb1f7620ee8ce33c8dc4674c4ce0596\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5blv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9tdbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:19Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:19 crc kubenswrapper[4967]: I1008 00:07:19.228846 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f1af2ba-8907-4d14-afa1-ffde207c80f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4535cffb8c9d64e662e282ecb44f6817b4604866f0bb75015f3b888d5554f821\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b03b83ddff429dbfa8fe052901aaa86a39ff0dd75477ffa8a6b6f4198f5819ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c9fa0559eb35ba67e11e36069ac8d0b2fec76399788eef668a08ad4353df3ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3312263e12bd42fbed0476ab5ace66b5497f62c9477865e130d67e3922e05cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e69dc603055aa221d44d11c049ddcc50cec568979911a310ed773a90341a499\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115f7a33f3dedae8f59d2f45f465f95f48b99c824e895ef7a0be9fe7a1b61d23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9972551ade22d86d9ccd63097a793a2fb4ce74779a99f4b9aab1cf215087975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9972551ade22d86d9ccd63097a793a2fb4ce74779a99f4b9aab1cf215087975\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T00:07:15Z\\\",\\\"message\\\":\\\"ring admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:15Z is after 2025-08-24T17:21:41Z]\\\\nI1008 00:07:15.446035 6404 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-marketplace/marketplace-operator-metrics]} name:Service_openshift-marketplace/marketplace-operator-metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.53:8081: 10.217.5.53:8383:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {89fe421e-04e8-4967-ac75-77a0e6f784ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1008 00:07:15.446063 6404 obj_retry.go:303] Retry object s\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:14Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-bs8w9_openshift-ovn-kubernetes(4f1af2ba-8907-4d14-afa1-ffde207c80f4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27fb2b1225e386dcd7fecaaa0aa54ba1fdfafabfde6c3824ae07315313067542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bs8w9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:19Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:19 crc kubenswrapper[4967]: I1008 00:07:19.242147 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-v5n7t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abe7ae2d-7e63-4249-96e1-4000e9628ccc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52575e853ff6727dde7c89e4add4362be126127db27191c2a7bf2871efa719d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpsjd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:06Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-v5n7t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:19Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:19 crc kubenswrapper[4967]: I1008 00:07:19.290343 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:19 crc kubenswrapper[4967]: I1008 00:07:19.290395 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:19 crc kubenswrapper[4967]: I1008 00:07:19.290417 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:19 crc kubenswrapper[4967]: I1008 00:07:19.290454 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:19 crc kubenswrapper[4967]: I1008 00:07:19.290475 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:19Z","lastTransitionTime":"2025-10-08T00:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:19 crc kubenswrapper[4967]: I1008 00:07:19.393290 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:19 crc kubenswrapper[4967]: I1008 00:07:19.393349 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:19 crc kubenswrapper[4967]: I1008 00:07:19.393367 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:19 crc kubenswrapper[4967]: I1008 00:07:19.393391 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:19 crc kubenswrapper[4967]: I1008 00:07:19.393408 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:19Z","lastTransitionTime":"2025-10-08T00:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:19 crc kubenswrapper[4967]: I1008 00:07:19.495832 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:19 crc kubenswrapper[4967]: I1008 00:07:19.495893 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:19 crc kubenswrapper[4967]: I1008 00:07:19.495912 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:19 crc kubenswrapper[4967]: I1008 00:07:19.495935 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:19 crc kubenswrapper[4967]: I1008 00:07:19.495953 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:19Z","lastTransitionTime":"2025-10-08T00:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:19 crc kubenswrapper[4967]: I1008 00:07:19.599145 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:19 crc kubenswrapper[4967]: I1008 00:07:19.599455 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:19 crc kubenswrapper[4967]: I1008 00:07:19.599594 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:19 crc kubenswrapper[4967]: I1008 00:07:19.599790 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:19 crc kubenswrapper[4967]: I1008 00:07:19.599931 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:19Z","lastTransitionTime":"2025-10-08T00:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:19 crc kubenswrapper[4967]: I1008 00:07:19.667047 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/59f68771-e25d-4367-87ee-68a31d767882-metrics-certs\") pod \"network-metrics-daemon-2h7z8\" (UID: \"59f68771-e25d-4367-87ee-68a31d767882\") " pod="openshift-multus/network-metrics-daemon-2h7z8" Oct 08 00:07:19 crc kubenswrapper[4967]: E1008 00:07:19.667308 4967 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 00:07:19 crc kubenswrapper[4967]: E1008 00:07:19.667494 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/59f68771-e25d-4367-87ee-68a31d767882-metrics-certs podName:59f68771-e25d-4367-87ee-68a31d767882 nodeName:}" failed. No retries permitted until 2025-10-08 00:07:20.667454469 +0000 UTC m=+39.317989168 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/59f68771-e25d-4367-87ee-68a31d767882-metrics-certs") pod "network-metrics-daemon-2h7z8" (UID: "59f68771-e25d-4367-87ee-68a31d767882") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 00:07:19 crc kubenswrapper[4967]: I1008 00:07:19.703541 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:19 crc kubenswrapper[4967]: I1008 00:07:19.703620 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:19 crc kubenswrapper[4967]: I1008 00:07:19.703709 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:19 crc kubenswrapper[4967]: I1008 00:07:19.703742 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:19 crc kubenswrapper[4967]: I1008 00:07:19.703760 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:19Z","lastTransitionTime":"2025-10-08T00:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:19 crc kubenswrapper[4967]: I1008 00:07:19.806756 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:19 crc kubenswrapper[4967]: I1008 00:07:19.806816 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:19 crc kubenswrapper[4967]: I1008 00:07:19.806832 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:19 crc kubenswrapper[4967]: I1008 00:07:19.806853 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:19 crc kubenswrapper[4967]: I1008 00:07:19.806866 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:19Z","lastTransitionTime":"2025-10-08T00:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:19 crc kubenswrapper[4967]: I1008 00:07:19.910152 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:19 crc kubenswrapper[4967]: I1008 00:07:19.910515 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:19 crc kubenswrapper[4967]: I1008 00:07:19.910728 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:19 crc kubenswrapper[4967]: I1008 00:07:19.910915 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:19 crc kubenswrapper[4967]: I1008 00:07:19.911045 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:19Z","lastTransitionTime":"2025-10-08T00:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:20 crc kubenswrapper[4967]: I1008 00:07:20.014152 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:20 crc kubenswrapper[4967]: I1008 00:07:20.014199 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:20 crc kubenswrapper[4967]: I1008 00:07:20.014219 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:20 crc kubenswrapper[4967]: I1008 00:07:20.014243 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:20 crc kubenswrapper[4967]: I1008 00:07:20.014260 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:20Z","lastTransitionTime":"2025-10-08T00:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:20 crc kubenswrapper[4967]: I1008 00:07:20.117062 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2h7z8" Oct 08 00:07:20 crc kubenswrapper[4967]: I1008 00:07:20.117204 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 00:07:20 crc kubenswrapper[4967]: E1008 00:07:20.117277 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2h7z8" podUID="59f68771-e25d-4367-87ee-68a31d767882" Oct 08 00:07:20 crc kubenswrapper[4967]: E1008 00:07:20.117355 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 00:07:20 crc kubenswrapper[4967]: I1008 00:07:20.117601 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:20 crc kubenswrapper[4967]: I1008 00:07:20.117651 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:20 crc kubenswrapper[4967]: I1008 00:07:20.117697 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:20 crc kubenswrapper[4967]: I1008 00:07:20.117723 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:20 crc kubenswrapper[4967]: I1008 00:07:20.117739 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:20Z","lastTransitionTime":"2025-10-08T00:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:20 crc kubenswrapper[4967]: I1008 00:07:20.220791 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:20 crc kubenswrapper[4967]: I1008 00:07:20.220844 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:20 crc kubenswrapper[4967]: I1008 00:07:20.220862 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:20 crc kubenswrapper[4967]: I1008 00:07:20.220884 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:20 crc kubenswrapper[4967]: I1008 00:07:20.220901 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:20Z","lastTransitionTime":"2025-10-08T00:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:20 crc kubenswrapper[4967]: I1008 00:07:20.324146 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:20 crc kubenswrapper[4967]: I1008 00:07:20.324208 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:20 crc kubenswrapper[4967]: I1008 00:07:20.324224 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:20 crc kubenswrapper[4967]: I1008 00:07:20.324247 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:20 crc kubenswrapper[4967]: I1008 00:07:20.324263 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:20Z","lastTransitionTime":"2025-10-08T00:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:20 crc kubenswrapper[4967]: I1008 00:07:20.427713 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:20 crc kubenswrapper[4967]: I1008 00:07:20.427774 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:20 crc kubenswrapper[4967]: I1008 00:07:20.427800 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:20 crc kubenswrapper[4967]: I1008 00:07:20.427828 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:20 crc kubenswrapper[4967]: I1008 00:07:20.427849 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:20Z","lastTransitionTime":"2025-10-08T00:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:20 crc kubenswrapper[4967]: I1008 00:07:20.535120 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:20 crc kubenswrapper[4967]: I1008 00:07:20.535534 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:20 crc kubenswrapper[4967]: I1008 00:07:20.535746 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:20 crc kubenswrapper[4967]: I1008 00:07:20.536013 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:20 crc kubenswrapper[4967]: I1008 00:07:20.536156 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:20Z","lastTransitionTime":"2025-10-08T00:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:20 crc kubenswrapper[4967]: I1008 00:07:20.639397 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:20 crc kubenswrapper[4967]: I1008 00:07:20.639796 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:20 crc kubenswrapper[4967]: I1008 00:07:20.639959 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:20 crc kubenswrapper[4967]: I1008 00:07:20.640106 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:20 crc kubenswrapper[4967]: I1008 00:07:20.640294 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:20Z","lastTransitionTime":"2025-10-08T00:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:20 crc kubenswrapper[4967]: I1008 00:07:20.681443 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/59f68771-e25d-4367-87ee-68a31d767882-metrics-certs\") pod \"network-metrics-daemon-2h7z8\" (UID: \"59f68771-e25d-4367-87ee-68a31d767882\") " pod="openshift-multus/network-metrics-daemon-2h7z8" Oct 08 00:07:20 crc kubenswrapper[4967]: E1008 00:07:20.681640 4967 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 00:07:20 crc kubenswrapper[4967]: E1008 00:07:20.681774 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/59f68771-e25d-4367-87ee-68a31d767882-metrics-certs podName:59f68771-e25d-4367-87ee-68a31d767882 nodeName:}" failed. No retries permitted until 2025-10-08 00:07:22.681733409 +0000 UTC m=+41.332268088 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/59f68771-e25d-4367-87ee-68a31d767882-metrics-certs") pod "network-metrics-daemon-2h7z8" (UID: "59f68771-e25d-4367-87ee-68a31d767882") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 00:07:20 crc kubenswrapper[4967]: I1008 00:07:20.742740 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:20 crc kubenswrapper[4967]: I1008 00:07:20.742786 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:20 crc kubenswrapper[4967]: I1008 00:07:20.742801 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:20 crc kubenswrapper[4967]: I1008 00:07:20.742835 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:20 crc kubenswrapper[4967]: I1008 00:07:20.742847 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:20Z","lastTransitionTime":"2025-10-08T00:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:20 crc kubenswrapper[4967]: I1008 00:07:20.845180 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:20 crc kubenswrapper[4967]: I1008 00:07:20.845224 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:20 crc kubenswrapper[4967]: I1008 00:07:20.845237 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:20 crc kubenswrapper[4967]: I1008 00:07:20.845255 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:20 crc kubenswrapper[4967]: I1008 00:07:20.845267 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:20Z","lastTransitionTime":"2025-10-08T00:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:20 crc kubenswrapper[4967]: I1008 00:07:20.947915 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:20 crc kubenswrapper[4967]: I1008 00:07:20.947992 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:20 crc kubenswrapper[4967]: I1008 00:07:20.948018 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:20 crc kubenswrapper[4967]: I1008 00:07:20.948048 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:20 crc kubenswrapper[4967]: I1008 00:07:20.948073 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:20Z","lastTransitionTime":"2025-10-08T00:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:21 crc kubenswrapper[4967]: I1008 00:07:21.050986 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:21 crc kubenswrapper[4967]: I1008 00:07:21.051051 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:21 crc kubenswrapper[4967]: I1008 00:07:21.051069 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:21 crc kubenswrapper[4967]: I1008 00:07:21.051092 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:21 crc kubenswrapper[4967]: I1008 00:07:21.051109 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:21Z","lastTransitionTime":"2025-10-08T00:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:21 crc kubenswrapper[4967]: I1008 00:07:21.117171 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 00:07:21 crc kubenswrapper[4967]: E1008 00:07:21.117380 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 00:07:21 crc kubenswrapper[4967]: I1008 00:07:21.117174 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 00:07:21 crc kubenswrapper[4967]: E1008 00:07:21.117556 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 00:07:21 crc kubenswrapper[4967]: I1008 00:07:21.153940 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:21 crc kubenswrapper[4967]: I1008 00:07:21.154016 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:21 crc kubenswrapper[4967]: I1008 00:07:21.154042 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:21 crc kubenswrapper[4967]: I1008 00:07:21.154074 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:21 crc kubenswrapper[4967]: I1008 00:07:21.154097 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:21Z","lastTransitionTime":"2025-10-08T00:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:21 crc kubenswrapper[4967]: I1008 00:07:21.257300 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:21 crc kubenswrapper[4967]: I1008 00:07:21.257783 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:21 crc kubenswrapper[4967]: I1008 00:07:21.257980 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:21 crc kubenswrapper[4967]: I1008 00:07:21.258199 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:21 crc kubenswrapper[4967]: I1008 00:07:21.258380 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:21Z","lastTransitionTime":"2025-10-08T00:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:21 crc kubenswrapper[4967]: I1008 00:07:21.362108 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:21 crc kubenswrapper[4967]: I1008 00:07:21.362172 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:21 crc kubenswrapper[4967]: I1008 00:07:21.362189 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:21 crc kubenswrapper[4967]: I1008 00:07:21.362215 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:21 crc kubenswrapper[4967]: I1008 00:07:21.362234 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:21Z","lastTransitionTime":"2025-10-08T00:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:21 crc kubenswrapper[4967]: I1008 00:07:21.465906 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:21 crc kubenswrapper[4967]: I1008 00:07:21.465971 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:21 crc kubenswrapper[4967]: I1008 00:07:21.465992 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:21 crc kubenswrapper[4967]: I1008 00:07:21.466017 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:21 crc kubenswrapper[4967]: I1008 00:07:21.466034 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:21Z","lastTransitionTime":"2025-10-08T00:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:21 crc kubenswrapper[4967]: I1008 00:07:21.569336 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:21 crc kubenswrapper[4967]: I1008 00:07:21.569434 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:21 crc kubenswrapper[4967]: I1008 00:07:21.569455 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:21 crc kubenswrapper[4967]: I1008 00:07:21.569478 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:21 crc kubenswrapper[4967]: I1008 00:07:21.569494 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:21Z","lastTransitionTime":"2025-10-08T00:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:21 crc kubenswrapper[4967]: I1008 00:07:21.672844 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:21 crc kubenswrapper[4967]: I1008 00:07:21.672923 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:21 crc kubenswrapper[4967]: I1008 00:07:21.672947 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:21 crc kubenswrapper[4967]: I1008 00:07:21.672977 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:21 crc kubenswrapper[4967]: I1008 00:07:21.672997 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:21Z","lastTransitionTime":"2025-10-08T00:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:21 crc kubenswrapper[4967]: I1008 00:07:21.776232 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:21 crc kubenswrapper[4967]: I1008 00:07:21.776605 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:21 crc kubenswrapper[4967]: I1008 00:07:21.776829 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:21 crc kubenswrapper[4967]: I1008 00:07:21.777001 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:21 crc kubenswrapper[4967]: I1008 00:07:21.777131 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:21Z","lastTransitionTime":"2025-10-08T00:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:21 crc kubenswrapper[4967]: I1008 00:07:21.880883 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:21 crc kubenswrapper[4967]: I1008 00:07:21.880948 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:21 crc kubenswrapper[4967]: I1008 00:07:21.880966 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:21 crc kubenswrapper[4967]: I1008 00:07:21.880990 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:21 crc kubenswrapper[4967]: I1008 00:07:21.881006 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:21Z","lastTransitionTime":"2025-10-08T00:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:21 crc kubenswrapper[4967]: I1008 00:07:21.984512 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:21 crc kubenswrapper[4967]: I1008 00:07:21.984578 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:21 crc kubenswrapper[4967]: I1008 00:07:21.984600 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:21 crc kubenswrapper[4967]: I1008 00:07:21.984626 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:21 crc kubenswrapper[4967]: I1008 00:07:21.984644 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:21Z","lastTransitionTime":"2025-10-08T00:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:22 crc kubenswrapper[4967]: I1008 00:07:22.088001 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:22 crc kubenswrapper[4967]: I1008 00:07:22.088064 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:22 crc kubenswrapper[4967]: I1008 00:07:22.088083 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:22 crc kubenswrapper[4967]: I1008 00:07:22.088110 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:22 crc kubenswrapper[4967]: I1008 00:07:22.088128 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:22Z","lastTransitionTime":"2025-10-08T00:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:22 crc kubenswrapper[4967]: I1008 00:07:22.117915 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 00:07:22 crc kubenswrapper[4967]: E1008 00:07:22.118120 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 00:07:22 crc kubenswrapper[4967]: I1008 00:07:22.118274 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2h7z8" Oct 08 00:07:22 crc kubenswrapper[4967]: E1008 00:07:22.118542 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2h7z8" podUID="59f68771-e25d-4367-87ee-68a31d767882" Oct 08 00:07:22 crc kubenswrapper[4967]: I1008 00:07:22.136108 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9tdbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41264a12-9486-4fee-a08c-607b590e5a70\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://806e85df796ae300d279f73a0cc4433a5bb1f7620ee8ce33c8dc4674c4ce0596\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5blv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9tdbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:22Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:22 crc kubenswrapper[4967]: I1008 00:07:22.165407 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f1af2ba-8907-4d14-afa1-ffde207c80f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4535cffb8c9d64e662e282ecb44f6817b4604866f0bb75015f3b888d5554f821\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b03b83ddff429dbfa8fe052901aaa86a39ff0dd75477ffa8a6b6f4198f5819ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c9fa0559eb35ba67e11e36069ac8d0b2fec76399788eef668a08ad4353df3ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3312263e12bd42fbed0476ab5ace66b5497f62c9477865e130d67e3922e05cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e69dc603055aa221d44d11c049ddcc50cec568979911a310ed773a90341a499\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115f7a33f3dedae8f59d2f45f465f95f48b99c824e895ef7a0be9fe7a1b61d23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f9972551ade22d86d9ccd63097a793a2fb4ce74779a99f4b9aab1cf215087975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9972551ade22d86d9ccd63097a793a2fb4ce74779a99f4b9aab1cf215087975\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T00:07:15Z\\\",\\\"message\\\":\\\"ring admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:15Z is after 2025-08-24T17:21:41Z]\\\\nI1008 00:07:15.446035 6404 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-marketplace/marketplace-operator-metrics]} name:Service_openshift-marketplace/marketplace-operator-metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.53:8081: 10.217.5.53:8383:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {89fe421e-04e8-4967-ac75-77a0e6f784ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1008 00:07:15.446063 6404 obj_retry.go:303] Retry object s\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:14Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-bs8w9_openshift-ovn-kubernetes(4f1af2ba-8907-4d14-afa1-ffde207c80f4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27fb2b1225e386dcd7fecaaa0aa54ba1fdfafabfde6c3824ae07315313067542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bs8w9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:22Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:22 crc kubenswrapper[4967]: I1008 00:07:22.183497 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-v5n7t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abe7ae2d-7e63-4249-96e1-4000e9628ccc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52575e853ff6727dde7c89e4add4362be126127db27191c2a7bf2871efa719d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpsjd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:06Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-v5n7t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:22Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:22 crc kubenswrapper[4967]: I1008 00:07:22.191160 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:22 crc kubenswrapper[4967]: I1008 00:07:22.191228 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:22 crc kubenswrapper[4967]: I1008 00:07:22.191254 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:22 crc kubenswrapper[4967]: I1008 00:07:22.191284 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:22 crc kubenswrapper[4967]: I1008 00:07:22.191306 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:22Z","lastTransitionTime":"2025-10-08T00:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:22 crc kubenswrapper[4967]: I1008 00:07:22.204364 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94e63bc5641ce80d507f76a19b203f78c42368a7a4105f3fd09e19af21a1b735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:22Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:22 crc kubenswrapper[4967]: I1008 00:07:22.225263 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e98666fd3152bab57d3723e7da72aae2703b79f7ca73eb3c3e5f0e062a82592\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cf1cbe5548146fe359aafa351864125ec6b8bdd7b796aa4d311f6f3a16ad212\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:22Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:22 crc kubenswrapper[4967]: I1008 00:07:22.244021 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-npz2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c71501b1-aa9e-420e-b1bd-7e15365e7adc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dfe6c420d804b2d7daef4b5cfe3f87b77b369287c4d65553fa93f5f24f589d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gblk2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-npz2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:22Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:22 crc kubenswrapper[4967]: I1008 00:07:22.258505 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2h7z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59f68771-e25d-4367-87ee-68a31d767882\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7r2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7r2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:18Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2h7z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:22Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:22 crc kubenswrapper[4967]: I1008 00:07:22.276745 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:22Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:22 crc kubenswrapper[4967]: I1008 00:07:22.293169 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:22Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:22 crc kubenswrapper[4967]: I1008 00:07:22.294735 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:22 crc kubenswrapper[4967]: I1008 00:07:22.294803 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:22 crc kubenswrapper[4967]: I1008 00:07:22.294830 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:22 crc kubenswrapper[4967]: I1008 00:07:22.294861 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:22 crc kubenswrapper[4967]: I1008 00:07:22.294985 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:22Z","lastTransitionTime":"2025-10-08T00:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:22 crc kubenswrapper[4967]: I1008 00:07:22.314292 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9mrr7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9299bc4f-f8d3-4baf-9885-39573982fd64\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://994beafece9dead0d85ae74dc453a142460705fd66b52ba22f2fa799cf27da08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb49c8476b482416588c4d403016940730064a619c0641eb029c947a4bab1215\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb49c8476b482416588c4d403016940730064a619c0641eb029c947a4bab1215\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b7b63962dcbf477f1fa4d71826a4aa3358920a2c1b2f5e48716411b5bdca049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b7b63962dcbf477f1fa4d71826a4aa3358920a2c1b2f5e48716411b5bdca049\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1e770dc737173fd4f3220ed6c52b422ccbd829722386a37029e1c3e10d80e1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1e770dc737173fd4f3220ed6c52b422ccbd829722386a37029e1c3e10d80e1a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e357d970878c996f219a79f1876b363176c9847b9575f33090f62bc299529a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e357d970878c996f219a79f1876b363176c9847b9575f33090f62bc299529a95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f098966bfad49d06278eaa72be042596d488115e597737d9a0a0cfc78f99f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49f098966bfad49d06278eaa72be042596d488115e597737d9a0a0cfc78f99f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1d53876b7b863c612d597fd716c3260e8857b6542802073bb730f6756071b46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c1d53876b7b863c612d597fd716c3260e8857b6542802073bb730f6756071b46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9mrr7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:22Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:22 crc kubenswrapper[4967]: I1008 00:07:22.330771 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kpmlg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f975d47d-42a6-422e-80e2-d1ee282817fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a317ae6640eb38b3aee0f9738cca890e8ab3d92075b620cd312f40a0628d46c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qznqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ef61854aea40924c5372b0553cd0b7e8faa39a1d42fec5b65cf867c368f72c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qznqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-kpmlg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:22Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:22 crc kubenswrapper[4967]: I1008 00:07:22.350636 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58c4704-a88a-43b4-b944-4de626f13f7e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22f5dabb39da0a7bd01db4dae1437b9a231a1fcc0dfe749a4a5a637a72cf33c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01a2b3dc785520c79eae998482e5a1347725c76584357073238fe60a619edb77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6aa6f5e81feceab3d1c7587152a70546b365e629a69267e4d3084324a1de764a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63c2b7c4b2791bd1f16888c58b878dbff1f87884e68c9949de914d0d3e1d39fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:22Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:22 crc kubenswrapper[4967]: I1008 00:07:22.363345 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:22Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:22 crc kubenswrapper[4967]: I1008 00:07:22.386177 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbfcdc882e2be1cb9cf666c9e7775671078f7dbbfe8421dd15db737b89920a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:22Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:22 crc kubenswrapper[4967]: I1008 00:07:22.398520 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:22 crc kubenswrapper[4967]: I1008 00:07:22.398575 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:22 crc kubenswrapper[4967]: I1008 00:07:22.398593 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:22 crc kubenswrapper[4967]: I1008 00:07:22.398615 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:22 crc kubenswrapper[4967]: I1008 00:07:22.398633 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:22Z","lastTransitionTime":"2025-10-08T00:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:22 crc kubenswrapper[4967]: I1008 00:07:22.403384 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9461162a-c27d-465e-9ca7-2d3936877cd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9778d78ddc7325abd4d77bc5e71b0b114f50be5dda4a438c7cb6b05640c8a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z9tp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28760af04ef321d0f91b2bce3ad074d44b63dabf676cc37f3be6aa8726341a2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z9tp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xvdtp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:22Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:22 crc kubenswrapper[4967]: I1008 00:07:22.425101 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fa4b747-ab6e-4a11-9903-ef2dcb9668c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbadcdade274b6d9be940f22bc949e3c55cadf95dcad43c8cb7a1085c0e6708c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d0a0b2ef4a98ee828688081a74f4b7ecc31c8fae415d7257a560ecd43960ce3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96b53b1b283025b3d078a3f505d469d1872ad4bb8ab8e612c36c70b0b00a920\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef86a2422616e1e0d5c3afe844818c2bdf2150f941d1aefc506995f58587b2bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1400e110997c64fa33d6b8672c7e91cab5f782952ddf8e1dea2a882689026e7f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 00:06:56.760088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 00:06:56.761210 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-369805205/tls.crt::/tmp/serving-cert-369805205/tls.key\\\\\\\"\\\\nI1008 00:07:02.321448 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 00:07:02.325932 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 00:07:02.325980 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 00:07:02.326025 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 00:07:02.326042 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 00:07:02.336203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 00:07:02.336240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 00:07:02.336245 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 00:07:02.336250 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 00:07:02.336253 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 00:07:02.336256 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 00:07:02.336258 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 00:07:02.336430 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 00:07:02.340774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:46Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca28d8db19036e47a7ef3dadb70941f6c6cf5735e7703977d036185c46fce40\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5b222ad84c31d8746dfe03124c7287b401deb3ca2c62ce2b4fbd40e638cee0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5b222ad84c31d8746dfe03124c7287b401deb3ca2c62ce2b4fbd40e638cee0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:22Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:22 crc kubenswrapper[4967]: I1008 00:07:22.501270 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:22 crc kubenswrapper[4967]: I1008 00:07:22.501324 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:22 crc kubenswrapper[4967]: I1008 00:07:22.501341 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:22 crc kubenswrapper[4967]: I1008 00:07:22.501365 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:22 crc kubenswrapper[4967]: I1008 00:07:22.501384 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:22Z","lastTransitionTime":"2025-10-08T00:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:22 crc kubenswrapper[4967]: I1008 00:07:22.604605 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:22 crc kubenswrapper[4967]: I1008 00:07:22.604743 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:22 crc kubenswrapper[4967]: I1008 00:07:22.604760 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:22 crc kubenswrapper[4967]: I1008 00:07:22.604781 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:22 crc kubenswrapper[4967]: I1008 00:07:22.604794 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:22Z","lastTransitionTime":"2025-10-08T00:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:22 crc kubenswrapper[4967]: I1008 00:07:22.704491 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/59f68771-e25d-4367-87ee-68a31d767882-metrics-certs\") pod \"network-metrics-daemon-2h7z8\" (UID: \"59f68771-e25d-4367-87ee-68a31d767882\") " pod="openshift-multus/network-metrics-daemon-2h7z8" Oct 08 00:07:22 crc kubenswrapper[4967]: E1008 00:07:22.704777 4967 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 00:07:22 crc kubenswrapper[4967]: E1008 00:07:22.704885 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/59f68771-e25d-4367-87ee-68a31d767882-metrics-certs podName:59f68771-e25d-4367-87ee-68a31d767882 nodeName:}" failed. No retries permitted until 2025-10-08 00:07:26.704862356 +0000 UTC m=+45.355397015 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/59f68771-e25d-4367-87ee-68a31d767882-metrics-certs") pod "network-metrics-daemon-2h7z8" (UID: "59f68771-e25d-4367-87ee-68a31d767882") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 00:07:22 crc kubenswrapper[4967]: I1008 00:07:22.707823 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:22 crc kubenswrapper[4967]: I1008 00:07:22.708031 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:22 crc kubenswrapper[4967]: I1008 00:07:22.708181 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:22 crc kubenswrapper[4967]: I1008 00:07:22.708458 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:22 crc kubenswrapper[4967]: I1008 00:07:22.708594 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:22Z","lastTransitionTime":"2025-10-08T00:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:22 crc kubenswrapper[4967]: I1008 00:07:22.811341 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:22 crc kubenswrapper[4967]: I1008 00:07:22.811754 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:22 crc kubenswrapper[4967]: I1008 00:07:22.811956 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:22 crc kubenswrapper[4967]: I1008 00:07:22.812145 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:22 crc kubenswrapper[4967]: I1008 00:07:22.812309 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:22Z","lastTransitionTime":"2025-10-08T00:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:22 crc kubenswrapper[4967]: I1008 00:07:22.915606 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:22 crc kubenswrapper[4967]: I1008 00:07:22.915650 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:22 crc kubenswrapper[4967]: I1008 00:07:22.915663 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:22 crc kubenswrapper[4967]: I1008 00:07:22.915711 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:22 crc kubenswrapper[4967]: I1008 00:07:22.915728 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:22Z","lastTransitionTime":"2025-10-08T00:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:23 crc kubenswrapper[4967]: I1008 00:07:23.019173 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:23 crc kubenswrapper[4967]: I1008 00:07:23.019207 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:23 crc kubenswrapper[4967]: I1008 00:07:23.019218 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:23 crc kubenswrapper[4967]: I1008 00:07:23.019233 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:23 crc kubenswrapper[4967]: I1008 00:07:23.019241 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:23Z","lastTransitionTime":"2025-10-08T00:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:23 crc kubenswrapper[4967]: I1008 00:07:23.117023 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 00:07:23 crc kubenswrapper[4967]: I1008 00:07:23.117084 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 00:07:23 crc kubenswrapper[4967]: E1008 00:07:23.117218 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 00:07:23 crc kubenswrapper[4967]: E1008 00:07:23.117374 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 00:07:23 crc kubenswrapper[4967]: I1008 00:07:23.123940 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:23 crc kubenswrapper[4967]: I1008 00:07:23.123989 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:23 crc kubenswrapper[4967]: I1008 00:07:23.124010 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:23 crc kubenswrapper[4967]: I1008 00:07:23.124036 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:23 crc kubenswrapper[4967]: I1008 00:07:23.124058 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:23Z","lastTransitionTime":"2025-10-08T00:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:23 crc kubenswrapper[4967]: I1008 00:07:23.227087 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:23 crc kubenswrapper[4967]: I1008 00:07:23.227191 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:23 crc kubenswrapper[4967]: I1008 00:07:23.227211 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:23 crc kubenswrapper[4967]: I1008 00:07:23.227275 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:23 crc kubenswrapper[4967]: I1008 00:07:23.227298 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:23Z","lastTransitionTime":"2025-10-08T00:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:23 crc kubenswrapper[4967]: I1008 00:07:23.330538 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:23 crc kubenswrapper[4967]: I1008 00:07:23.330603 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:23 crc kubenswrapper[4967]: I1008 00:07:23.330621 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:23 crc kubenswrapper[4967]: I1008 00:07:23.330648 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:23 crc kubenswrapper[4967]: I1008 00:07:23.330698 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:23Z","lastTransitionTime":"2025-10-08T00:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:23 crc kubenswrapper[4967]: I1008 00:07:23.433277 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:23 crc kubenswrapper[4967]: I1008 00:07:23.433338 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:23 crc kubenswrapper[4967]: I1008 00:07:23.433353 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:23 crc kubenswrapper[4967]: I1008 00:07:23.433374 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:23 crc kubenswrapper[4967]: I1008 00:07:23.433392 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:23Z","lastTransitionTime":"2025-10-08T00:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:23 crc kubenswrapper[4967]: I1008 00:07:23.536054 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:23 crc kubenswrapper[4967]: I1008 00:07:23.536090 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:23 crc kubenswrapper[4967]: I1008 00:07:23.536100 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:23 crc kubenswrapper[4967]: I1008 00:07:23.536113 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:23 crc kubenswrapper[4967]: I1008 00:07:23.536121 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:23Z","lastTransitionTime":"2025-10-08T00:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:23 crc kubenswrapper[4967]: I1008 00:07:23.638511 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:23 crc kubenswrapper[4967]: I1008 00:07:23.638545 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:23 crc kubenswrapper[4967]: I1008 00:07:23.638559 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:23 crc kubenswrapper[4967]: I1008 00:07:23.638577 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:23 crc kubenswrapper[4967]: I1008 00:07:23.638589 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:23Z","lastTransitionTime":"2025-10-08T00:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:23 crc kubenswrapper[4967]: I1008 00:07:23.740586 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:23 crc kubenswrapper[4967]: I1008 00:07:23.740651 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:23 crc kubenswrapper[4967]: I1008 00:07:23.740707 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:23 crc kubenswrapper[4967]: I1008 00:07:23.740732 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:23 crc kubenswrapper[4967]: I1008 00:07:23.740749 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:23Z","lastTransitionTime":"2025-10-08T00:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:23 crc kubenswrapper[4967]: I1008 00:07:23.843491 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:23 crc kubenswrapper[4967]: I1008 00:07:23.843558 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:23 crc kubenswrapper[4967]: I1008 00:07:23.843575 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:23 crc kubenswrapper[4967]: I1008 00:07:23.843603 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:23 crc kubenswrapper[4967]: I1008 00:07:23.843624 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:23Z","lastTransitionTime":"2025-10-08T00:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:23 crc kubenswrapper[4967]: I1008 00:07:23.946626 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:23 crc kubenswrapper[4967]: I1008 00:07:23.946724 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:23 crc kubenswrapper[4967]: I1008 00:07:23.946743 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:23 crc kubenswrapper[4967]: I1008 00:07:23.946761 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:23 crc kubenswrapper[4967]: I1008 00:07:23.946774 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:23Z","lastTransitionTime":"2025-10-08T00:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:24 crc kubenswrapper[4967]: I1008 00:07:24.049501 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:24 crc kubenswrapper[4967]: I1008 00:07:24.049598 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:24 crc kubenswrapper[4967]: I1008 00:07:24.049618 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:24 crc kubenswrapper[4967]: I1008 00:07:24.049638 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:24 crc kubenswrapper[4967]: I1008 00:07:24.049655 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:24Z","lastTransitionTime":"2025-10-08T00:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:24 crc kubenswrapper[4967]: I1008 00:07:24.117954 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2h7z8" Oct 08 00:07:24 crc kubenswrapper[4967]: I1008 00:07:24.117962 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 00:07:24 crc kubenswrapper[4967]: E1008 00:07:24.118246 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2h7z8" podUID="59f68771-e25d-4367-87ee-68a31d767882" Oct 08 00:07:24 crc kubenswrapper[4967]: E1008 00:07:24.118353 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 00:07:24 crc kubenswrapper[4967]: I1008 00:07:24.153062 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:24 crc kubenswrapper[4967]: I1008 00:07:24.153177 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:24 crc kubenswrapper[4967]: I1008 00:07:24.153231 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:24 crc kubenswrapper[4967]: I1008 00:07:24.153259 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:24 crc kubenswrapper[4967]: I1008 00:07:24.153316 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:24Z","lastTransitionTime":"2025-10-08T00:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:24 crc kubenswrapper[4967]: I1008 00:07:24.256626 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:24 crc kubenswrapper[4967]: I1008 00:07:24.256709 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:24 crc kubenswrapper[4967]: I1008 00:07:24.256722 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:24 crc kubenswrapper[4967]: I1008 00:07:24.256741 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:24 crc kubenswrapper[4967]: I1008 00:07:24.256755 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:24Z","lastTransitionTime":"2025-10-08T00:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:24 crc kubenswrapper[4967]: I1008 00:07:24.359456 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:24 crc kubenswrapper[4967]: I1008 00:07:24.359549 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:24 crc kubenswrapper[4967]: I1008 00:07:24.359599 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:24 crc kubenswrapper[4967]: I1008 00:07:24.359623 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:24 crc kubenswrapper[4967]: I1008 00:07:24.359776 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:24Z","lastTransitionTime":"2025-10-08T00:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:24 crc kubenswrapper[4967]: I1008 00:07:24.462333 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:24 crc kubenswrapper[4967]: I1008 00:07:24.462380 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:24 crc kubenswrapper[4967]: I1008 00:07:24.462391 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:24 crc kubenswrapper[4967]: I1008 00:07:24.462407 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:24 crc kubenswrapper[4967]: I1008 00:07:24.462421 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:24Z","lastTransitionTime":"2025-10-08T00:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:24 crc kubenswrapper[4967]: I1008 00:07:24.565801 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:24 crc kubenswrapper[4967]: I1008 00:07:24.565871 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:24 crc kubenswrapper[4967]: I1008 00:07:24.565889 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:24 crc kubenswrapper[4967]: I1008 00:07:24.565913 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:24 crc kubenswrapper[4967]: I1008 00:07:24.565932 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:24Z","lastTransitionTime":"2025-10-08T00:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:24 crc kubenswrapper[4967]: I1008 00:07:24.669560 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:24 crc kubenswrapper[4967]: I1008 00:07:24.669626 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:24 crc kubenswrapper[4967]: I1008 00:07:24.669644 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:24 crc kubenswrapper[4967]: I1008 00:07:24.669697 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:24 crc kubenswrapper[4967]: I1008 00:07:24.669715 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:24Z","lastTransitionTime":"2025-10-08T00:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:24 crc kubenswrapper[4967]: I1008 00:07:24.772464 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:24 crc kubenswrapper[4967]: I1008 00:07:24.772528 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:24 crc kubenswrapper[4967]: I1008 00:07:24.772546 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:24 crc kubenswrapper[4967]: I1008 00:07:24.772569 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:24 crc kubenswrapper[4967]: I1008 00:07:24.772588 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:24Z","lastTransitionTime":"2025-10-08T00:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:24 crc kubenswrapper[4967]: I1008 00:07:24.874791 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:24 crc kubenswrapper[4967]: I1008 00:07:24.874836 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:24 crc kubenswrapper[4967]: I1008 00:07:24.874846 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:24 crc kubenswrapper[4967]: I1008 00:07:24.874859 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:24 crc kubenswrapper[4967]: I1008 00:07:24.874870 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:24Z","lastTransitionTime":"2025-10-08T00:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:24 crc kubenswrapper[4967]: I1008 00:07:24.977501 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:24 crc kubenswrapper[4967]: I1008 00:07:24.977543 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:24 crc kubenswrapper[4967]: I1008 00:07:24.977552 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:24 crc kubenswrapper[4967]: I1008 00:07:24.977568 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:24 crc kubenswrapper[4967]: I1008 00:07:24.977577 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:24Z","lastTransitionTime":"2025-10-08T00:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:25 crc kubenswrapper[4967]: I1008 00:07:25.080333 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:25 crc kubenswrapper[4967]: I1008 00:07:25.080383 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:25 crc kubenswrapper[4967]: I1008 00:07:25.080402 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:25 crc kubenswrapper[4967]: I1008 00:07:25.080425 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:25 crc kubenswrapper[4967]: I1008 00:07:25.080441 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:25Z","lastTransitionTime":"2025-10-08T00:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:25 crc kubenswrapper[4967]: I1008 00:07:25.117475 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 00:07:25 crc kubenswrapper[4967]: I1008 00:07:25.117586 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 00:07:25 crc kubenswrapper[4967]: E1008 00:07:25.117727 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 00:07:25 crc kubenswrapper[4967]: E1008 00:07:25.117949 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 00:07:25 crc kubenswrapper[4967]: I1008 00:07:25.183552 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:25 crc kubenswrapper[4967]: I1008 00:07:25.183613 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:25 crc kubenswrapper[4967]: I1008 00:07:25.183631 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:25 crc kubenswrapper[4967]: I1008 00:07:25.183652 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:25 crc kubenswrapper[4967]: I1008 00:07:25.183698 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:25Z","lastTransitionTime":"2025-10-08T00:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:25 crc kubenswrapper[4967]: I1008 00:07:25.286237 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:25 crc kubenswrapper[4967]: I1008 00:07:25.286305 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:25 crc kubenswrapper[4967]: I1008 00:07:25.286316 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:25 crc kubenswrapper[4967]: I1008 00:07:25.286332 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:25 crc kubenswrapper[4967]: I1008 00:07:25.286341 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:25Z","lastTransitionTime":"2025-10-08T00:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:25 crc kubenswrapper[4967]: I1008 00:07:25.388596 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:25 crc kubenswrapper[4967]: I1008 00:07:25.388699 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:25 crc kubenswrapper[4967]: I1008 00:07:25.388715 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:25 crc kubenswrapper[4967]: I1008 00:07:25.388731 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:25 crc kubenswrapper[4967]: I1008 00:07:25.388740 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:25Z","lastTransitionTime":"2025-10-08T00:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:25 crc kubenswrapper[4967]: I1008 00:07:25.491377 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:25 crc kubenswrapper[4967]: I1008 00:07:25.491415 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:25 crc kubenswrapper[4967]: I1008 00:07:25.491428 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:25 crc kubenswrapper[4967]: I1008 00:07:25.491446 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:25 crc kubenswrapper[4967]: I1008 00:07:25.491458 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:25Z","lastTransitionTime":"2025-10-08T00:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:25 crc kubenswrapper[4967]: I1008 00:07:25.594211 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:25 crc kubenswrapper[4967]: I1008 00:07:25.594245 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:25 crc kubenswrapper[4967]: I1008 00:07:25.594255 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:25 crc kubenswrapper[4967]: I1008 00:07:25.594272 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:25 crc kubenswrapper[4967]: I1008 00:07:25.594285 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:25Z","lastTransitionTime":"2025-10-08T00:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:25 crc kubenswrapper[4967]: I1008 00:07:25.697152 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:25 crc kubenswrapper[4967]: I1008 00:07:25.697192 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:25 crc kubenswrapper[4967]: I1008 00:07:25.697218 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:25 crc kubenswrapper[4967]: I1008 00:07:25.697233 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:25 crc kubenswrapper[4967]: I1008 00:07:25.697242 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:25Z","lastTransitionTime":"2025-10-08T00:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:25 crc kubenswrapper[4967]: I1008 00:07:25.799463 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:25 crc kubenswrapper[4967]: I1008 00:07:25.799840 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:25 crc kubenswrapper[4967]: I1008 00:07:25.799953 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:25 crc kubenswrapper[4967]: I1008 00:07:25.800069 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:25 crc kubenswrapper[4967]: I1008 00:07:25.800149 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:25Z","lastTransitionTime":"2025-10-08T00:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:25 crc kubenswrapper[4967]: I1008 00:07:25.903481 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:25 crc kubenswrapper[4967]: I1008 00:07:25.903820 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:25 crc kubenswrapper[4967]: I1008 00:07:25.904001 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:25 crc kubenswrapper[4967]: I1008 00:07:25.904193 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:25 crc kubenswrapper[4967]: I1008 00:07:25.904312 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:25Z","lastTransitionTime":"2025-10-08T00:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:26 crc kubenswrapper[4967]: I1008 00:07:26.007430 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:26 crc kubenswrapper[4967]: I1008 00:07:26.007463 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:26 crc kubenswrapper[4967]: I1008 00:07:26.007473 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:26 crc kubenswrapper[4967]: I1008 00:07:26.007485 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:26 crc kubenswrapper[4967]: I1008 00:07:26.007495 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:26Z","lastTransitionTime":"2025-10-08T00:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:26 crc kubenswrapper[4967]: I1008 00:07:26.110897 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:26 crc kubenswrapper[4967]: I1008 00:07:26.111195 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:26 crc kubenswrapper[4967]: I1008 00:07:26.111324 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:26 crc kubenswrapper[4967]: I1008 00:07:26.111423 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:26 crc kubenswrapper[4967]: I1008 00:07:26.111507 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:26Z","lastTransitionTime":"2025-10-08T00:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:26 crc kubenswrapper[4967]: I1008 00:07:26.117428 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 00:07:26 crc kubenswrapper[4967]: E1008 00:07:26.117533 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 00:07:26 crc kubenswrapper[4967]: I1008 00:07:26.117428 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2h7z8" Oct 08 00:07:26 crc kubenswrapper[4967]: E1008 00:07:26.117648 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2h7z8" podUID="59f68771-e25d-4367-87ee-68a31d767882" Oct 08 00:07:26 crc kubenswrapper[4967]: I1008 00:07:26.213875 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:26 crc kubenswrapper[4967]: I1008 00:07:26.213930 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:26 crc kubenswrapper[4967]: I1008 00:07:26.213943 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:26 crc kubenswrapper[4967]: I1008 00:07:26.213959 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:26 crc kubenswrapper[4967]: I1008 00:07:26.213971 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:26Z","lastTransitionTime":"2025-10-08T00:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:26 crc kubenswrapper[4967]: I1008 00:07:26.317030 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:26 crc kubenswrapper[4967]: I1008 00:07:26.317122 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:26 crc kubenswrapper[4967]: I1008 00:07:26.317150 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:26 crc kubenswrapper[4967]: I1008 00:07:26.317188 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:26 crc kubenswrapper[4967]: I1008 00:07:26.317213 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:26Z","lastTransitionTime":"2025-10-08T00:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:26 crc kubenswrapper[4967]: I1008 00:07:26.420120 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:26 crc kubenswrapper[4967]: I1008 00:07:26.420164 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:26 crc kubenswrapper[4967]: I1008 00:07:26.420174 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:26 crc kubenswrapper[4967]: I1008 00:07:26.420188 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:26 crc kubenswrapper[4967]: I1008 00:07:26.420197 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:26Z","lastTransitionTime":"2025-10-08T00:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:26 crc kubenswrapper[4967]: I1008 00:07:26.522042 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:26 crc kubenswrapper[4967]: I1008 00:07:26.522108 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:26 crc kubenswrapper[4967]: I1008 00:07:26.522118 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:26 crc kubenswrapper[4967]: I1008 00:07:26.522131 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:26 crc kubenswrapper[4967]: I1008 00:07:26.522140 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:26Z","lastTransitionTime":"2025-10-08T00:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:26 crc kubenswrapper[4967]: I1008 00:07:26.625480 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:26 crc kubenswrapper[4967]: I1008 00:07:26.625557 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:26 crc kubenswrapper[4967]: I1008 00:07:26.625592 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:26 crc kubenswrapper[4967]: I1008 00:07:26.625622 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:26 crc kubenswrapper[4967]: I1008 00:07:26.625648 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:26Z","lastTransitionTime":"2025-10-08T00:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:26 crc kubenswrapper[4967]: I1008 00:07:26.728623 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:26 crc kubenswrapper[4967]: I1008 00:07:26.728734 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:26 crc kubenswrapper[4967]: I1008 00:07:26.728755 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:26 crc kubenswrapper[4967]: I1008 00:07:26.728781 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:26 crc kubenswrapper[4967]: I1008 00:07:26.728798 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:26Z","lastTransitionTime":"2025-10-08T00:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:26 crc kubenswrapper[4967]: I1008 00:07:26.749186 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/59f68771-e25d-4367-87ee-68a31d767882-metrics-certs\") pod \"network-metrics-daemon-2h7z8\" (UID: \"59f68771-e25d-4367-87ee-68a31d767882\") " pod="openshift-multus/network-metrics-daemon-2h7z8" Oct 08 00:07:26 crc kubenswrapper[4967]: E1008 00:07:26.749351 4967 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 00:07:26 crc kubenswrapper[4967]: E1008 00:07:26.749412 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/59f68771-e25d-4367-87ee-68a31d767882-metrics-certs podName:59f68771-e25d-4367-87ee-68a31d767882 nodeName:}" failed. No retries permitted until 2025-10-08 00:07:34.749395558 +0000 UTC m=+53.399930207 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/59f68771-e25d-4367-87ee-68a31d767882-metrics-certs") pod "network-metrics-daemon-2h7z8" (UID: "59f68771-e25d-4367-87ee-68a31d767882") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 00:07:26 crc kubenswrapper[4967]: I1008 00:07:26.831416 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:26 crc kubenswrapper[4967]: I1008 00:07:26.831452 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:26 crc kubenswrapper[4967]: I1008 00:07:26.831461 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:26 crc kubenswrapper[4967]: I1008 00:07:26.831476 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:26 crc kubenswrapper[4967]: I1008 00:07:26.831490 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:26Z","lastTransitionTime":"2025-10-08T00:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:26 crc kubenswrapper[4967]: I1008 00:07:26.934761 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:26 crc kubenswrapper[4967]: I1008 00:07:26.934830 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:26 crc kubenswrapper[4967]: I1008 00:07:26.934853 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:26 crc kubenswrapper[4967]: I1008 00:07:26.934885 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:26 crc kubenswrapper[4967]: I1008 00:07:26.934907 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:26Z","lastTransitionTime":"2025-10-08T00:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:27 crc kubenswrapper[4967]: I1008 00:07:27.037269 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:27 crc kubenswrapper[4967]: I1008 00:07:27.037310 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:27 crc kubenswrapper[4967]: I1008 00:07:27.037323 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:27 crc kubenswrapper[4967]: I1008 00:07:27.037341 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:27 crc kubenswrapper[4967]: I1008 00:07:27.037354 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:27Z","lastTransitionTime":"2025-10-08T00:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:27 crc kubenswrapper[4967]: I1008 00:07:27.117217 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 00:07:27 crc kubenswrapper[4967]: I1008 00:07:27.117374 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 00:07:27 crc kubenswrapper[4967]: E1008 00:07:27.117631 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 00:07:27 crc kubenswrapper[4967]: E1008 00:07:27.117867 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 00:07:27 crc kubenswrapper[4967]: I1008 00:07:27.140640 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:27 crc kubenswrapper[4967]: I1008 00:07:27.140696 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:27 crc kubenswrapper[4967]: I1008 00:07:27.140707 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:27 crc kubenswrapper[4967]: I1008 00:07:27.140722 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:27 crc kubenswrapper[4967]: I1008 00:07:27.140731 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:27Z","lastTransitionTime":"2025-10-08T00:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:27 crc kubenswrapper[4967]: I1008 00:07:27.243474 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:27 crc kubenswrapper[4967]: I1008 00:07:27.243515 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:27 crc kubenswrapper[4967]: I1008 00:07:27.243527 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:27 crc kubenswrapper[4967]: I1008 00:07:27.243542 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:27 crc kubenswrapper[4967]: I1008 00:07:27.243551 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:27Z","lastTransitionTime":"2025-10-08T00:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:27 crc kubenswrapper[4967]: I1008 00:07:27.347498 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:27 crc kubenswrapper[4967]: I1008 00:07:27.347541 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:27 crc kubenswrapper[4967]: I1008 00:07:27.347550 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:27 crc kubenswrapper[4967]: I1008 00:07:27.347565 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:27 crc kubenswrapper[4967]: I1008 00:07:27.347578 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:27Z","lastTransitionTime":"2025-10-08T00:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:27 crc kubenswrapper[4967]: I1008 00:07:27.450815 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:27 crc kubenswrapper[4967]: I1008 00:07:27.450865 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:27 crc kubenswrapper[4967]: I1008 00:07:27.450878 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:27 crc kubenswrapper[4967]: I1008 00:07:27.450897 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:27 crc kubenswrapper[4967]: I1008 00:07:27.450908 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:27Z","lastTransitionTime":"2025-10-08T00:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:27 crc kubenswrapper[4967]: I1008 00:07:27.555958 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:27 crc kubenswrapper[4967]: I1008 00:07:27.555997 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:27 crc kubenswrapper[4967]: I1008 00:07:27.556006 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:27 crc kubenswrapper[4967]: I1008 00:07:27.556020 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:27 crc kubenswrapper[4967]: I1008 00:07:27.556029 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:27Z","lastTransitionTime":"2025-10-08T00:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:27 crc kubenswrapper[4967]: I1008 00:07:27.658322 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:27 crc kubenswrapper[4967]: I1008 00:07:27.658364 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:27 crc kubenswrapper[4967]: I1008 00:07:27.658377 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:27 crc kubenswrapper[4967]: I1008 00:07:27.658394 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:27 crc kubenswrapper[4967]: I1008 00:07:27.658403 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:27Z","lastTransitionTime":"2025-10-08T00:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:27 crc kubenswrapper[4967]: I1008 00:07:27.760808 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:27 crc kubenswrapper[4967]: I1008 00:07:27.760865 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:27 crc kubenswrapper[4967]: I1008 00:07:27.760875 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:27 crc kubenswrapper[4967]: I1008 00:07:27.760891 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:27 crc kubenswrapper[4967]: I1008 00:07:27.760903 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:27Z","lastTransitionTime":"2025-10-08T00:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:27 crc kubenswrapper[4967]: I1008 00:07:27.863077 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:27 crc kubenswrapper[4967]: I1008 00:07:27.863116 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:27 crc kubenswrapper[4967]: I1008 00:07:27.863129 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:27 crc kubenswrapper[4967]: I1008 00:07:27.863144 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:27 crc kubenswrapper[4967]: I1008 00:07:27.863154 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:27Z","lastTransitionTime":"2025-10-08T00:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:27 crc kubenswrapper[4967]: I1008 00:07:27.965540 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:27 crc kubenswrapper[4967]: I1008 00:07:27.965585 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:27 crc kubenswrapper[4967]: I1008 00:07:27.965604 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:27 crc kubenswrapper[4967]: I1008 00:07:27.965619 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:27 crc kubenswrapper[4967]: I1008 00:07:27.965629 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:27Z","lastTransitionTime":"2025-10-08T00:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:28 crc kubenswrapper[4967]: I1008 00:07:28.068835 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:28 crc kubenswrapper[4967]: I1008 00:07:28.068886 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:28 crc kubenswrapper[4967]: I1008 00:07:28.068895 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:28 crc kubenswrapper[4967]: I1008 00:07:28.068914 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:28 crc kubenswrapper[4967]: I1008 00:07:28.068924 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:28Z","lastTransitionTime":"2025-10-08T00:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:28 crc kubenswrapper[4967]: I1008 00:07:28.116917 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 00:07:28 crc kubenswrapper[4967]: I1008 00:07:28.117011 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2h7z8" Oct 08 00:07:28 crc kubenswrapper[4967]: E1008 00:07:28.117099 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 00:07:28 crc kubenswrapper[4967]: E1008 00:07:28.117267 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2h7z8" podUID="59f68771-e25d-4367-87ee-68a31d767882" Oct 08 00:07:28 crc kubenswrapper[4967]: I1008 00:07:28.117890 4967 scope.go:117] "RemoveContainer" containerID="f9972551ade22d86d9ccd63097a793a2fb4ce74779a99f4b9aab1cf215087975" Oct 08 00:07:28 crc kubenswrapper[4967]: I1008 00:07:28.172742 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:28 crc kubenswrapper[4967]: I1008 00:07:28.173140 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:28 crc kubenswrapper[4967]: I1008 00:07:28.173181 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:28 crc kubenswrapper[4967]: I1008 00:07:28.173203 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:28 crc kubenswrapper[4967]: I1008 00:07:28.173216 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:28Z","lastTransitionTime":"2025-10-08T00:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:28 crc kubenswrapper[4967]: I1008 00:07:28.276746 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:28 crc kubenswrapper[4967]: I1008 00:07:28.276791 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:28 crc kubenswrapper[4967]: I1008 00:07:28.276809 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:28 crc kubenswrapper[4967]: I1008 00:07:28.276836 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:28 crc kubenswrapper[4967]: I1008 00:07:28.276854 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:28Z","lastTransitionTime":"2025-10-08T00:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:28 crc kubenswrapper[4967]: I1008 00:07:28.379395 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:28 crc kubenswrapper[4967]: I1008 00:07:28.379430 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:28 crc kubenswrapper[4967]: I1008 00:07:28.379452 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:28 crc kubenswrapper[4967]: I1008 00:07:28.379468 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:28 crc kubenswrapper[4967]: I1008 00:07:28.379479 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:28Z","lastTransitionTime":"2025-10-08T00:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:28 crc kubenswrapper[4967]: I1008 00:07:28.482024 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:28 crc kubenswrapper[4967]: I1008 00:07:28.482082 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:28 crc kubenswrapper[4967]: I1008 00:07:28.482107 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:28 crc kubenswrapper[4967]: I1008 00:07:28.482130 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:28 crc kubenswrapper[4967]: I1008 00:07:28.482146 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:28Z","lastTransitionTime":"2025-10-08T00:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:28 crc kubenswrapper[4967]: I1008 00:07:28.528077 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bs8w9_4f1af2ba-8907-4d14-afa1-ffde207c80f4/ovnkube-controller/1.log" Oct 08 00:07:28 crc kubenswrapper[4967]: I1008 00:07:28.530838 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" event={"ID":"4f1af2ba-8907-4d14-afa1-ffde207c80f4","Type":"ContainerStarted","Data":"d1cfd26b02d0e17b111b8ad84dab6c2adf692de66ff77158ee7cbff941c66ecd"} Oct 08 00:07:28 crc kubenswrapper[4967]: I1008 00:07:28.530969 4967 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 08 00:07:28 crc kubenswrapper[4967]: I1008 00:07:28.551462 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94e63bc5641ce80d507f76a19b203f78c42368a7a4105f3fd09e19af21a1b735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:28Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:28 crc kubenswrapper[4967]: I1008 00:07:28.570905 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e98666fd3152bab57d3723e7da72aae2703b79f7ca73eb3c3e5f0e062a82592\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cf1cbe5548146fe359aafa351864125ec6b8bdd7b796aa4d311f6f3a16ad212\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:28Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:28 crc kubenswrapper[4967]: I1008 00:07:28.584379 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:28 crc kubenswrapper[4967]: I1008 00:07:28.584420 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:28 crc kubenswrapper[4967]: I1008 00:07:28.584431 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:28 crc kubenswrapper[4967]: I1008 00:07:28.584447 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:28 crc kubenswrapper[4967]: I1008 00:07:28.584458 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:28Z","lastTransitionTime":"2025-10-08T00:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:28 crc kubenswrapper[4967]: I1008 00:07:28.589230 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9tdbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41264a12-9486-4fee-a08c-607b590e5a70\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://806e85df796ae300d279f73a0cc4433a5bb1f7620ee8ce33c8dc4674c4ce0596\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5blv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9tdbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:28Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:28 crc kubenswrapper[4967]: I1008 00:07:28.609033 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f1af2ba-8907-4d14-afa1-ffde207c80f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4535cffb8c9d64e662e282ecb44f6817b4604866f0bb75015f3b888d5554f821\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b03b83ddff429dbfa8fe052901aaa86a39ff0dd75477ffa8a6b6f4198f5819ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c9fa0559eb35ba67e11e36069ac8d0b2fec76399788eef668a08ad4353df3ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3312263e12bd42fbed0476ab5ace66b5497f62c9477865e130d67e3922e05cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e69dc603055aa221d44d11c049ddcc50cec568979911a310ed773a90341a499\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115f7a33f3dedae8f59d2f45f465f95f48b99c824e895ef7a0be9fe7a1b61d23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1cfd26b02d0e17b111b8ad84dab6c2adf692de66ff77158ee7cbff941c66ecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9972551ade22d86d9ccd63097a793a2fb4ce74779a99f4b9aab1cf215087975\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T00:07:15Z\\\",\\\"message\\\":\\\"ring admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:15Z is after 2025-08-24T17:21:41Z]\\\\nI1008 00:07:15.446035 6404 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-marketplace/marketplace-operator-metrics]} name:Service_openshift-marketplace/marketplace-operator-metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.53:8081: 10.217.5.53:8383:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {89fe421e-04e8-4967-ac75-77a0e6f784ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1008 00:07:15.446063 6404 obj_retry.go:303] Retry object s\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:14Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27fb2b1225e386dcd7fecaaa0aa54ba1fdfafabfde6c3824ae07315313067542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bs8w9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:28Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:28 crc kubenswrapper[4967]: I1008 00:07:28.619028 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-v5n7t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abe7ae2d-7e63-4249-96e1-4000e9628ccc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52575e853ff6727dde7c89e4add4362be126127db27191c2a7bf2871efa719d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpsjd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:06Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-v5n7t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:28Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:28 crc kubenswrapper[4967]: I1008 00:07:28.630374 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:28Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:28 crc kubenswrapper[4967]: I1008 00:07:28.643139 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-npz2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c71501b1-aa9e-420e-b1bd-7e15365e7adc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dfe6c420d804b2d7daef4b5cfe3f87b77b369287c4d65553fa93f5f24f589d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gblk2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-npz2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:28Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:28 crc kubenswrapper[4967]: I1008 00:07:28.651976 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2h7z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59f68771-e25d-4367-87ee-68a31d767882\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7r2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7r2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:18Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2h7z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:28Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:28 crc kubenswrapper[4967]: I1008 00:07:28.663825 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58c4704-a88a-43b4-b944-4de626f13f7e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22f5dabb39da0a7bd01db4dae1437b9a231a1fcc0dfe749a4a5a637a72cf33c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01a2b3dc785520c79eae998482e5a1347725c76584357073238fe60a619edb77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6aa6f5e81feceab3d1c7587152a70546b365e629a69267e4d3084324a1de764a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63c2b7c4b2791bd1f16888c58b878dbff1f87884e68c9949de914d0d3e1d39fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:28Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:28 crc kubenswrapper[4967]: I1008 00:07:28.687225 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:28 crc kubenswrapper[4967]: I1008 00:07:28.687286 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:28 crc kubenswrapper[4967]: I1008 00:07:28.687299 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:28 crc kubenswrapper[4967]: I1008 00:07:28.687317 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:28 crc kubenswrapper[4967]: I1008 00:07:28.687330 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:28Z","lastTransitionTime":"2025-10-08T00:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:28 crc kubenswrapper[4967]: I1008 00:07:28.693511 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:28Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:28 crc kubenswrapper[4967]: I1008 00:07:28.711036 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9mrr7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9299bc4f-f8d3-4baf-9885-39573982fd64\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://994beafece9dead0d85ae74dc453a142460705fd66b52ba22f2fa799cf27da08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb49c8476b482416588c4d403016940730064a619c0641eb029c947a4bab1215\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb49c8476b482416588c4d403016940730064a619c0641eb029c947a4bab1215\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b7b63962dcbf477f1fa4d71826a4aa3358920a2c1b2f5e48716411b5bdca049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b7b63962dcbf477f1fa4d71826a4aa3358920a2c1b2f5e48716411b5bdca049\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1e770dc737173fd4f3220ed6c52b422ccbd829722386a37029e1c3e10d80e1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1e770dc737173fd4f3220ed6c52b422ccbd829722386a37029e1c3e10d80e1a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e357d970878c996f219a79f1876b363176c9847b9575f33090f62bc299529a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e357d970878c996f219a79f1876b363176c9847b9575f33090f62bc299529a95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f098966bfad49d06278eaa72be042596d488115e597737d9a0a0cfc78f99f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49f098966bfad49d06278eaa72be042596d488115e597737d9a0a0cfc78f99f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1d53876b7b863c612d597fd716c3260e8857b6542802073bb730f6756071b46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c1d53876b7b863c612d597fd716c3260e8857b6542802073bb730f6756071b46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9mrr7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:28Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:28 crc kubenswrapper[4967]: I1008 00:07:28.722449 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kpmlg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f975d47d-42a6-422e-80e2-d1ee282817fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a317ae6640eb38b3aee0f9738cca890e8ab3d92075b620cd312f40a0628d46c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qznqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ef61854aea40924c5372b0553cd0b7e8faa39a1d42fec5b65cf867c368f72c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qznqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-kpmlg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:28Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:28 crc kubenswrapper[4967]: I1008 00:07:28.736338 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fa4b747-ab6e-4a11-9903-ef2dcb9668c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbadcdade274b6d9be940f22bc949e3c55cadf95dcad43c8cb7a1085c0e6708c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d0a0b2ef4a98ee828688081a74f4b7ecc31c8fae415d7257a560ecd43960ce3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96b53b1b283025b3d078a3f505d469d1872ad4bb8ab8e612c36c70b0b00a920\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef86a2422616e1e0d5c3afe844818c2bdf2150f941d1aefc506995f58587b2bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1400e110997c64fa33d6b8672c7e91cab5f782952ddf8e1dea2a882689026e7f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 00:06:56.760088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 00:06:56.761210 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-369805205/tls.crt::/tmp/serving-cert-369805205/tls.key\\\\\\\"\\\\nI1008 00:07:02.321448 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 00:07:02.325932 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 00:07:02.325980 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 00:07:02.326025 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 00:07:02.326042 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 00:07:02.336203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 00:07:02.336240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 00:07:02.336245 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 00:07:02.336250 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 00:07:02.336253 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 00:07:02.336256 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 00:07:02.336258 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 00:07:02.336430 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 00:07:02.340774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:46Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca28d8db19036e47a7ef3dadb70941f6c6cf5735e7703977d036185c46fce40\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5b222ad84c31d8746dfe03124c7287b401deb3ca2c62ce2b4fbd40e638cee0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5b222ad84c31d8746dfe03124c7287b401deb3ca2c62ce2b4fbd40e638cee0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:28Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:28 crc kubenswrapper[4967]: I1008 00:07:28.749476 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:28Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:28 crc kubenswrapper[4967]: I1008 00:07:28.762462 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbfcdc882e2be1cb9cf666c9e7775671078f7dbbfe8421dd15db737b89920a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:28Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:28 crc kubenswrapper[4967]: I1008 00:07:28.776988 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9461162a-c27d-465e-9ca7-2d3936877cd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9778d78ddc7325abd4d77bc5e71b0b114f50be5dda4a438c7cb6b05640c8a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z9tp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28760af04ef321d0f91b2bce3ad074d44b63dabf676cc37f3be6aa8726341a2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z9tp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xvdtp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:28Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:28 crc kubenswrapper[4967]: I1008 00:07:28.790738 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:28 crc kubenswrapper[4967]: I1008 00:07:28.790784 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:28 crc kubenswrapper[4967]: I1008 00:07:28.790796 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:28 crc kubenswrapper[4967]: I1008 00:07:28.790814 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:28 crc kubenswrapper[4967]: I1008 00:07:28.790826 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:28Z","lastTransitionTime":"2025-10-08T00:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:28 crc kubenswrapper[4967]: I1008 00:07:28.893193 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:28 crc kubenswrapper[4967]: I1008 00:07:28.893242 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:28 crc kubenswrapper[4967]: I1008 00:07:28.893255 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:28 crc kubenswrapper[4967]: I1008 00:07:28.893274 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:28 crc kubenswrapper[4967]: I1008 00:07:28.893285 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:28Z","lastTransitionTime":"2025-10-08T00:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.002706 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.002771 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.002787 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.002809 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.002824 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:29Z","lastTransitionTime":"2025-10-08T00:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.105489 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.105550 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.105564 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.105584 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.105596 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:29Z","lastTransitionTime":"2025-10-08T00:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.117503 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 00:07:29 crc kubenswrapper[4967]: E1008 00:07:29.117739 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.118017 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 00:07:29 crc kubenswrapper[4967]: E1008 00:07:29.118136 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.208080 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.208132 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.208146 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.208164 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.208180 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:29Z","lastTransitionTime":"2025-10-08T00:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.307989 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.308057 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.308073 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.308093 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.308111 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:29Z","lastTransitionTime":"2025-10-08T00:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:29 crc kubenswrapper[4967]: E1008 00:07:29.325004 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5cddad4c-22d1-42d7-a373-87b47c1cc97b\\\",\\\"systemUUID\\\":\\\"b243ca1c-20cd-4e6f-96b2-52fce74dd2d4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:29Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.330179 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.330360 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.330453 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.330549 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.330638 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:29Z","lastTransitionTime":"2025-10-08T00:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:29 crc kubenswrapper[4967]: E1008 00:07:29.350849 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5cddad4c-22d1-42d7-a373-87b47c1cc97b\\\",\\\"systemUUID\\\":\\\"b243ca1c-20cd-4e6f-96b2-52fce74dd2d4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:29Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.355156 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.355213 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.355230 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.355254 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.355269 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:29Z","lastTransitionTime":"2025-10-08T00:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:29 crc kubenswrapper[4967]: E1008 00:07:29.369514 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5cddad4c-22d1-42d7-a373-87b47c1cc97b\\\",\\\"systemUUID\\\":\\\"b243ca1c-20cd-4e6f-96b2-52fce74dd2d4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:29Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.373711 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.373758 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.373770 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.373796 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.373816 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:29Z","lastTransitionTime":"2025-10-08T00:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:29 crc kubenswrapper[4967]: E1008 00:07:29.393211 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5cddad4c-22d1-42d7-a373-87b47c1cc97b\\\",\\\"systemUUID\\\":\\\"b243ca1c-20cd-4e6f-96b2-52fce74dd2d4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:29Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.398341 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.398398 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.398425 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.398455 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.398470 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:29Z","lastTransitionTime":"2025-10-08T00:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:29 crc kubenswrapper[4967]: E1008 00:07:29.414466 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5cddad4c-22d1-42d7-a373-87b47c1cc97b\\\",\\\"systemUUID\\\":\\\"b243ca1c-20cd-4e6f-96b2-52fce74dd2d4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:29Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:29 crc kubenswrapper[4967]: E1008 00:07:29.414614 4967 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.416626 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.416731 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.416750 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.416774 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.416785 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:29Z","lastTransitionTime":"2025-10-08T00:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.519342 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.519619 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.519823 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.519998 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.520097 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:29Z","lastTransitionTime":"2025-10-08T00:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.536797 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bs8w9_4f1af2ba-8907-4d14-afa1-ffde207c80f4/ovnkube-controller/2.log" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.537392 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bs8w9_4f1af2ba-8907-4d14-afa1-ffde207c80f4/ovnkube-controller/1.log" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.541614 4967 generic.go:334] "Generic (PLEG): container finished" podID="4f1af2ba-8907-4d14-afa1-ffde207c80f4" containerID="d1cfd26b02d0e17b111b8ad84dab6c2adf692de66ff77158ee7cbff941c66ecd" exitCode=1 Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.541663 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" event={"ID":"4f1af2ba-8907-4d14-afa1-ffde207c80f4","Type":"ContainerDied","Data":"d1cfd26b02d0e17b111b8ad84dab6c2adf692de66ff77158ee7cbff941c66ecd"} Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.541726 4967 scope.go:117] "RemoveContainer" containerID="f9972551ade22d86d9ccd63097a793a2fb4ce74779a99f4b9aab1cf215087975" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.542961 4967 scope.go:117] "RemoveContainer" containerID="d1cfd26b02d0e17b111b8ad84dab6c2adf692de66ff77158ee7cbff941c66ecd" Oct 08 00:07:29 crc kubenswrapper[4967]: E1008 00:07:29.543212 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-bs8w9_openshift-ovn-kubernetes(4f1af2ba-8907-4d14-afa1-ffde207c80f4)\"" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" podUID="4f1af2ba-8907-4d14-afa1-ffde207c80f4" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.563368 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58c4704-a88a-43b4-b944-4de626f13f7e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22f5dabb39da0a7bd01db4dae1437b9a231a1fcc0dfe749a4a5a637a72cf33c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01a2b3dc785520c79eae998482e5a1347725c76584357073238fe60a619edb77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6aa6f5e81feceab3d1c7587152a70546b365e629a69267e4d3084324a1de764a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63c2b7c4b2791bd1f16888c58b878dbff1f87884e68c9949de914d0d3e1d39fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:29Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.585875 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:29Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.606424 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9mrr7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9299bc4f-f8d3-4baf-9885-39573982fd64\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://994beafece9dead0d85ae74dc453a142460705fd66b52ba22f2fa799cf27da08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb49c8476b482416588c4d403016940730064a619c0641eb029c947a4bab1215\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb49c8476b482416588c4d403016940730064a619c0641eb029c947a4bab1215\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b7b63962dcbf477f1fa4d71826a4aa3358920a2c1b2f5e48716411b5bdca049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b7b63962dcbf477f1fa4d71826a4aa3358920a2c1b2f5e48716411b5bdca049\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1e770dc737173fd4f3220ed6c52b422ccbd829722386a37029e1c3e10d80e1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1e770dc737173fd4f3220ed6c52b422ccbd829722386a37029e1c3e10d80e1a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e357d970878c996f219a79f1876b363176c9847b9575f33090f62bc299529a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e357d970878c996f219a79f1876b363176c9847b9575f33090f62bc299529a95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f098966bfad49d06278eaa72be042596d488115e597737d9a0a0cfc78f99f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49f098966bfad49d06278eaa72be042596d488115e597737d9a0a0cfc78f99f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1d53876b7b863c612d597fd716c3260e8857b6542802073bb730f6756071b46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c1d53876b7b863c612d597fd716c3260e8857b6542802073bb730f6756071b46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9mrr7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:29Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.621605 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kpmlg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f975d47d-42a6-422e-80e2-d1ee282817fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a317ae6640eb38b3aee0f9738cca890e8ab3d92075b620cd312f40a0628d46c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qznqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ef61854aea40924c5372b0553cd0b7e8faa39a1d42fec5b65cf867c368f72c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qznqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-kpmlg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:29Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.622546 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.622572 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.622585 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.622601 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.622613 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:29Z","lastTransitionTime":"2025-10-08T00:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.634569 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:29Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.645417 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbfcdc882e2be1cb9cf666c9e7775671078f7dbbfe8421dd15db737b89920a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:29Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.657476 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9461162a-c27d-465e-9ca7-2d3936877cd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9778d78ddc7325abd4d77bc5e71b0b114f50be5dda4a438c7cb6b05640c8a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z9tp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28760af04ef321d0f91b2bce3ad074d44b63dabf676cc37f3be6aa8726341a2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z9tp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xvdtp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:29Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.672186 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fa4b747-ab6e-4a11-9903-ef2dcb9668c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbadcdade274b6d9be940f22bc949e3c55cadf95dcad43c8cb7a1085c0e6708c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d0a0b2ef4a98ee828688081a74f4b7ecc31c8fae415d7257a560ecd43960ce3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96b53b1b283025b3d078a3f505d469d1872ad4bb8ab8e612c36c70b0b00a920\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef86a2422616e1e0d5c3afe844818c2bdf2150f941d1aefc506995f58587b2bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1400e110997c64fa33d6b8672c7e91cab5f782952ddf8e1dea2a882689026e7f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 00:06:56.760088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 00:06:56.761210 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-369805205/tls.crt::/tmp/serving-cert-369805205/tls.key\\\\\\\"\\\\nI1008 00:07:02.321448 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 00:07:02.325932 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 00:07:02.325980 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 00:07:02.326025 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 00:07:02.326042 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 00:07:02.336203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 00:07:02.336240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 00:07:02.336245 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 00:07:02.336250 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 00:07:02.336253 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 00:07:02.336256 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 00:07:02.336258 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 00:07:02.336430 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 00:07:02.340774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:46Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca28d8db19036e47a7ef3dadb70941f6c6cf5735e7703977d036185c46fce40\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5b222ad84c31d8746dfe03124c7287b401deb3ca2c62ce2b4fbd40e638cee0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5b222ad84c31d8746dfe03124c7287b401deb3ca2c62ce2b4fbd40e638cee0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:29Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.691157 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e98666fd3152bab57d3723e7da72aae2703b79f7ca73eb3c3e5f0e062a82592\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cf1cbe5548146fe359aafa351864125ec6b8bdd7b796aa4d311f6f3a16ad212\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:29Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.703217 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9tdbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41264a12-9486-4fee-a08c-607b590e5a70\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://806e85df796ae300d279f73a0cc4433a5bb1f7620ee8ce33c8dc4674c4ce0596\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5blv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9tdbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:29Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.726183 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.726234 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.726250 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.726270 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.726285 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:29Z","lastTransitionTime":"2025-10-08T00:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.729497 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f1af2ba-8907-4d14-afa1-ffde207c80f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4535cffb8c9d64e662e282ecb44f6817b4604866f0bb75015f3b888d5554f821\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b03b83ddff429dbfa8fe052901aaa86a39ff0dd75477ffa8a6b6f4198f5819ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c9fa0559eb35ba67e11e36069ac8d0b2fec76399788eef668a08ad4353df3ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3312263e12bd42fbed0476ab5ace66b5497f62c9477865e130d67e3922e05cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e69dc603055aa221d44d11c049ddcc50cec568979911a310ed773a90341a499\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115f7a33f3dedae8f59d2f45f465f95f48b99c824e895ef7a0be9fe7a1b61d23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1cfd26b02d0e17b111b8ad84dab6c2adf692de66ff77158ee7cbff941c66ecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f9972551ade22d86d9ccd63097a793a2fb4ce74779a99f4b9aab1cf215087975\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T00:07:15Z\\\",\\\"message\\\":\\\"ring admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:15Z is after 2025-08-24T17:21:41Z]\\\\nI1008 00:07:15.446035 6404 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-marketplace/marketplace-operator-metrics]} name:Service_openshift-marketplace/marketplace-operator-metrics_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.53:8081: 10.217.5.53:8383:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {89fe421e-04e8-4967-ac75-77a0e6f784ef}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1008 00:07:15.446063 6404 obj_retry.go:303] Retry object s\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:14Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1cfd26b02d0e17b111b8ad84dab6c2adf692de66ff77158ee7cbff941c66ecd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T00:07:29Z\\\",\\\"message\\\":\\\"iled objects of type *v1.Pod\\\\nI1008 00:07:28.973629 6638 obj_retry.go:409] Going to retry *v1.Pod resource setup for 1 objects: [openshift-multus/network-metrics-daemon-2h7z8]\\\\nI1008 00:07:28.973646 6638 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1008 00:07:28.973685 6638 obj_retry.go:285] Attempting retry of *v1.Pod openshift-multus/network-metrics-daemon-2h7z8 before timer (time: 2025-10-08 00:07:30.377116097 +0000 UTC m=+1.948919645): skip\\\\nI1008 00:07:28.973706 6638 obj_retry.go:420] Function iterateRetryResources for *v1.Pod ended (in 79.172µs)\\\\nI1008 00:07:28.973904 6638 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1008 00:07:28.973958 6638 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1008 00:07:28.974084 6638 factory.go:656] Stopping watch factory\\\\nI1008 00:07:28.974107 6638 ovnkube.go:599] Stopped ovnkube\\\\nI1008 00:07:28.974137 6638 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1008 00:07:28.974149 6638 handler.go:208] Removed *v1.Node event handler 2\\\\nI1008 00:07:28.974163 6638 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1008 00:07:28.974267 6638 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27fb2b1225e386dcd7fecaaa0aa54ba1fdfafabfde6c3824ae07315313067542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bs8w9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:29Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.743243 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-v5n7t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abe7ae2d-7e63-4249-96e1-4000e9628ccc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52575e853ff6727dde7c89e4add4362be126127db27191c2a7bf2871efa719d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpsjd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:06Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-v5n7t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:29Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.760968 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94e63bc5641ce80d507f76a19b203f78c42368a7a4105f3fd09e19af21a1b735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:29Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.773597 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:29Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.790014 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-npz2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c71501b1-aa9e-420e-b1bd-7e15365e7adc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dfe6c420d804b2d7daef4b5cfe3f87b77b369287c4d65553fa93f5f24f589d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gblk2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-npz2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:29Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.800538 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2h7z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59f68771-e25d-4367-87ee-68a31d767882\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7r2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7r2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:18Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2h7z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:29Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.828783 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.828836 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.828854 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.828875 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.828891 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:29Z","lastTransitionTime":"2025-10-08T00:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.872349 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.930817 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.931095 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.931157 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.931215 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:29 crc kubenswrapper[4967]: I1008 00:07:29.931271 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:29Z","lastTransitionTime":"2025-10-08T00:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:30 crc kubenswrapper[4967]: I1008 00:07:30.033686 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:30 crc kubenswrapper[4967]: I1008 00:07:30.033728 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:30 crc kubenswrapper[4967]: I1008 00:07:30.033742 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:30 crc kubenswrapper[4967]: I1008 00:07:30.033761 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:30 crc kubenswrapper[4967]: I1008 00:07:30.033777 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:30Z","lastTransitionTime":"2025-10-08T00:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:30 crc kubenswrapper[4967]: I1008 00:07:30.117274 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 00:07:30 crc kubenswrapper[4967]: I1008 00:07:30.117274 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2h7z8" Oct 08 00:07:30 crc kubenswrapper[4967]: E1008 00:07:30.117450 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 00:07:30 crc kubenswrapper[4967]: E1008 00:07:30.117658 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2h7z8" podUID="59f68771-e25d-4367-87ee-68a31d767882" Oct 08 00:07:30 crc kubenswrapper[4967]: I1008 00:07:30.135746 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:30 crc kubenswrapper[4967]: I1008 00:07:30.135802 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:30 crc kubenswrapper[4967]: I1008 00:07:30.135819 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:30 crc kubenswrapper[4967]: I1008 00:07:30.135842 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:30 crc kubenswrapper[4967]: I1008 00:07:30.135859 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:30Z","lastTransitionTime":"2025-10-08T00:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:30 crc kubenswrapper[4967]: I1008 00:07:30.238499 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:30 crc kubenswrapper[4967]: I1008 00:07:30.238817 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:30 crc kubenswrapper[4967]: I1008 00:07:30.238894 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:30 crc kubenswrapper[4967]: I1008 00:07:30.238966 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:30 crc kubenswrapper[4967]: I1008 00:07:30.239060 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:30Z","lastTransitionTime":"2025-10-08T00:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:30 crc kubenswrapper[4967]: I1008 00:07:30.341423 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:30 crc kubenswrapper[4967]: I1008 00:07:30.341454 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:30 crc kubenswrapper[4967]: I1008 00:07:30.341463 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:30 crc kubenswrapper[4967]: I1008 00:07:30.341475 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:30 crc kubenswrapper[4967]: I1008 00:07:30.341484 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:30Z","lastTransitionTime":"2025-10-08T00:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:30 crc kubenswrapper[4967]: I1008 00:07:30.443842 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:30 crc kubenswrapper[4967]: I1008 00:07:30.443903 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:30 crc kubenswrapper[4967]: I1008 00:07:30.443926 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:30 crc kubenswrapper[4967]: I1008 00:07:30.443950 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:30 crc kubenswrapper[4967]: I1008 00:07:30.443967 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:30Z","lastTransitionTime":"2025-10-08T00:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:30 crc kubenswrapper[4967]: I1008 00:07:30.546052 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:30 crc kubenswrapper[4967]: I1008 00:07:30.546092 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:30 crc kubenswrapper[4967]: I1008 00:07:30.546105 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:30 crc kubenswrapper[4967]: I1008 00:07:30.546121 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:30 crc kubenswrapper[4967]: I1008 00:07:30.546132 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:30Z","lastTransitionTime":"2025-10-08T00:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:30 crc kubenswrapper[4967]: I1008 00:07:30.547023 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bs8w9_4f1af2ba-8907-4d14-afa1-ffde207c80f4/ovnkube-controller/2.log" Oct 08 00:07:30 crc kubenswrapper[4967]: I1008 00:07:30.550648 4967 scope.go:117] "RemoveContainer" containerID="d1cfd26b02d0e17b111b8ad84dab6c2adf692de66ff77158ee7cbff941c66ecd" Oct 08 00:07:30 crc kubenswrapper[4967]: E1008 00:07:30.550891 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-bs8w9_openshift-ovn-kubernetes(4f1af2ba-8907-4d14-afa1-ffde207c80f4)\"" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" podUID="4f1af2ba-8907-4d14-afa1-ffde207c80f4" Oct 08 00:07:30 crc kubenswrapper[4967]: I1008 00:07:30.581257 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f1af2ba-8907-4d14-afa1-ffde207c80f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4535cffb8c9d64e662e282ecb44f6817b4604866f0bb75015f3b888d5554f821\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b03b83ddff429dbfa8fe052901aaa86a39ff0dd75477ffa8a6b6f4198f5819ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c9fa0559eb35ba67e11e36069ac8d0b2fec76399788eef668a08ad4353df3ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3312263e12bd42fbed0476ab5ace66b5497f62c9477865e130d67e3922e05cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e69dc603055aa221d44d11c049ddcc50cec568979911a310ed773a90341a499\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115f7a33f3dedae8f59d2f45f465f95f48b99c824e895ef7a0be9fe7a1b61d23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1cfd26b02d0e17b111b8ad84dab6c2adf692de66ff77158ee7cbff941c66ecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1cfd26b02d0e17b111b8ad84dab6c2adf692de66ff77158ee7cbff941c66ecd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T00:07:29Z\\\",\\\"message\\\":\\\"iled objects of type *v1.Pod\\\\nI1008 00:07:28.973629 6638 obj_retry.go:409] Going to retry *v1.Pod resource setup for 1 objects: [openshift-multus/network-metrics-daemon-2h7z8]\\\\nI1008 00:07:28.973646 6638 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1008 00:07:28.973685 6638 obj_retry.go:285] Attempting retry of *v1.Pod openshift-multus/network-metrics-daemon-2h7z8 before timer (time: 2025-10-08 00:07:30.377116097 +0000 UTC m=+1.948919645): skip\\\\nI1008 00:07:28.973706 6638 obj_retry.go:420] Function iterateRetryResources for *v1.Pod ended (in 79.172µs)\\\\nI1008 00:07:28.973904 6638 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1008 00:07:28.973958 6638 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1008 00:07:28.974084 6638 factory.go:656] Stopping watch factory\\\\nI1008 00:07:28.974107 6638 ovnkube.go:599] Stopped ovnkube\\\\nI1008 00:07:28.974137 6638 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1008 00:07:28.974149 6638 handler.go:208] Removed *v1.Node event handler 2\\\\nI1008 00:07:28.974163 6638 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1008 00:07:28.974267 6638 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-bs8w9_openshift-ovn-kubernetes(4f1af2ba-8907-4d14-afa1-ffde207c80f4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27fb2b1225e386dcd7fecaaa0aa54ba1fdfafabfde6c3824ae07315313067542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bs8w9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:30Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:30 crc kubenswrapper[4967]: I1008 00:07:30.595131 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-v5n7t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abe7ae2d-7e63-4249-96e1-4000e9628ccc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52575e853ff6727dde7c89e4add4362be126127db27191c2a7bf2871efa719d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpsjd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:06Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-v5n7t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:30Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:30 crc kubenswrapper[4967]: I1008 00:07:30.613895 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94e63bc5641ce80d507f76a19b203f78c42368a7a4105f3fd09e19af21a1b735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:30Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:30 crc kubenswrapper[4967]: I1008 00:07:30.636275 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e98666fd3152bab57d3723e7da72aae2703b79f7ca73eb3c3e5f0e062a82592\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cf1cbe5548146fe359aafa351864125ec6b8bdd7b796aa4d311f6f3a16ad212\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:30Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:30 crc kubenswrapper[4967]: I1008 00:07:30.649164 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:30 crc kubenswrapper[4967]: I1008 00:07:30.649256 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:30 crc kubenswrapper[4967]: I1008 00:07:30.649284 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:30 crc kubenswrapper[4967]: I1008 00:07:30.649309 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:30 crc kubenswrapper[4967]: I1008 00:07:30.649328 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:30Z","lastTransitionTime":"2025-10-08T00:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:30 crc kubenswrapper[4967]: I1008 00:07:30.649145 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9tdbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41264a12-9486-4fee-a08c-607b590e5a70\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://806e85df796ae300d279f73a0cc4433a5bb1f7620ee8ce33c8dc4674c4ce0596\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5blv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9tdbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:30Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:30 crc kubenswrapper[4967]: I1008 00:07:30.663616 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2h7z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59f68771-e25d-4367-87ee-68a31d767882\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7r2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7r2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:18Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2h7z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:30Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:30 crc kubenswrapper[4967]: I1008 00:07:30.677021 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:30Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:30 crc kubenswrapper[4967]: I1008 00:07:30.695636 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-npz2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c71501b1-aa9e-420e-b1bd-7e15365e7adc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dfe6c420d804b2d7daef4b5cfe3f87b77b369287c4d65553fa93f5f24f589d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gblk2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-npz2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:30Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:30 crc kubenswrapper[4967]: I1008 00:07:30.718872 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9mrr7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9299bc4f-f8d3-4baf-9885-39573982fd64\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://994beafece9dead0d85ae74dc453a142460705fd66b52ba22f2fa799cf27da08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb49c8476b482416588c4d403016940730064a619c0641eb029c947a4bab1215\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb49c8476b482416588c4d403016940730064a619c0641eb029c947a4bab1215\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b7b63962dcbf477f1fa4d71826a4aa3358920a2c1b2f5e48716411b5bdca049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b7b63962dcbf477f1fa4d71826a4aa3358920a2c1b2f5e48716411b5bdca049\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1e770dc737173fd4f3220ed6c52b422ccbd829722386a37029e1c3e10d80e1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1e770dc737173fd4f3220ed6c52b422ccbd829722386a37029e1c3e10d80e1a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e357d970878c996f219a79f1876b363176c9847b9575f33090f62bc299529a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e357d970878c996f219a79f1876b363176c9847b9575f33090f62bc299529a95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f098966bfad49d06278eaa72be042596d488115e597737d9a0a0cfc78f99f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49f098966bfad49d06278eaa72be042596d488115e597737d9a0a0cfc78f99f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1d53876b7b863c612d597fd716c3260e8857b6542802073bb730f6756071b46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c1d53876b7b863c612d597fd716c3260e8857b6542802073bb730f6756071b46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9mrr7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:30Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:30 crc kubenswrapper[4967]: I1008 00:07:30.739058 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kpmlg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f975d47d-42a6-422e-80e2-d1ee282817fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a317ae6640eb38b3aee0f9738cca890e8ab3d92075b620cd312f40a0628d46c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qznqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ef61854aea40924c5372b0553cd0b7e8faa39a1d42fec5b65cf867c368f72c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qznqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-kpmlg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:30Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:30 crc kubenswrapper[4967]: I1008 00:07:30.751663 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:30 crc kubenswrapper[4967]: I1008 00:07:30.751805 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:30 crc kubenswrapper[4967]: I1008 00:07:30.751816 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:30 crc kubenswrapper[4967]: I1008 00:07:30.751830 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:30 crc kubenswrapper[4967]: I1008 00:07:30.751839 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:30Z","lastTransitionTime":"2025-10-08T00:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:30 crc kubenswrapper[4967]: I1008 00:07:30.756562 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58c4704-a88a-43b4-b944-4de626f13f7e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22f5dabb39da0a7bd01db4dae1437b9a231a1fcc0dfe749a4a5a637a72cf33c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01a2b3dc785520c79eae998482e5a1347725c76584357073238fe60a619edb77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6aa6f5e81feceab3d1c7587152a70546b365e629a69267e4d3084324a1de764a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63c2b7c4b2791bd1f16888c58b878dbff1f87884e68c9949de914d0d3e1d39fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:30Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:30 crc kubenswrapper[4967]: I1008 00:07:30.769144 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:30Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:30 crc kubenswrapper[4967]: I1008 00:07:30.783513 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbfcdc882e2be1cb9cf666c9e7775671078f7dbbfe8421dd15db737b89920a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:30Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:30 crc kubenswrapper[4967]: I1008 00:07:30.798971 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9461162a-c27d-465e-9ca7-2d3936877cd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9778d78ddc7325abd4d77bc5e71b0b114f50be5dda4a438c7cb6b05640c8a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z9tp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28760af04ef321d0f91b2bce3ad074d44b63dabf676cc37f3be6aa8726341a2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z9tp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xvdtp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:30Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:30 crc kubenswrapper[4967]: I1008 00:07:30.813080 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fa4b747-ab6e-4a11-9903-ef2dcb9668c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbadcdade274b6d9be940f22bc949e3c55cadf95dcad43c8cb7a1085c0e6708c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d0a0b2ef4a98ee828688081a74f4b7ecc31c8fae415d7257a560ecd43960ce3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96b53b1b283025b3d078a3f505d469d1872ad4bb8ab8e612c36c70b0b00a920\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef86a2422616e1e0d5c3afe844818c2bdf2150f941d1aefc506995f58587b2bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1400e110997c64fa33d6b8672c7e91cab5f782952ddf8e1dea2a882689026e7f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 00:06:56.760088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 00:06:56.761210 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-369805205/tls.crt::/tmp/serving-cert-369805205/tls.key\\\\\\\"\\\\nI1008 00:07:02.321448 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 00:07:02.325932 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 00:07:02.325980 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 00:07:02.326025 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 00:07:02.326042 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 00:07:02.336203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 00:07:02.336240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 00:07:02.336245 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 00:07:02.336250 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 00:07:02.336253 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 00:07:02.336256 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 00:07:02.336258 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 00:07:02.336430 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 00:07:02.340774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:46Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca28d8db19036e47a7ef3dadb70941f6c6cf5735e7703977d036185c46fce40\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5b222ad84c31d8746dfe03124c7287b401deb3ca2c62ce2b4fbd40e638cee0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5b222ad84c31d8746dfe03124c7287b401deb3ca2c62ce2b4fbd40e638cee0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:30Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:30 crc kubenswrapper[4967]: I1008 00:07:30.825105 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:30Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:30 crc kubenswrapper[4967]: I1008 00:07:30.853646 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:30 crc kubenswrapper[4967]: I1008 00:07:30.853718 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:30 crc kubenswrapper[4967]: I1008 00:07:30.853730 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:30 crc kubenswrapper[4967]: I1008 00:07:30.853746 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:30 crc kubenswrapper[4967]: I1008 00:07:30.853758 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:30Z","lastTransitionTime":"2025-10-08T00:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:30 crc kubenswrapper[4967]: I1008 00:07:30.956190 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:30 crc kubenswrapper[4967]: I1008 00:07:30.956241 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:30 crc kubenswrapper[4967]: I1008 00:07:30.956258 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:30 crc kubenswrapper[4967]: I1008 00:07:30.956276 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:30 crc kubenswrapper[4967]: I1008 00:07:30.956319 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:30Z","lastTransitionTime":"2025-10-08T00:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:31 crc kubenswrapper[4967]: I1008 00:07:31.059771 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:31 crc kubenswrapper[4967]: I1008 00:07:31.059831 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:31 crc kubenswrapper[4967]: I1008 00:07:31.059849 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:31 crc kubenswrapper[4967]: I1008 00:07:31.059876 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:31 crc kubenswrapper[4967]: I1008 00:07:31.059895 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:31Z","lastTransitionTime":"2025-10-08T00:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:31 crc kubenswrapper[4967]: I1008 00:07:31.117422 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 00:07:31 crc kubenswrapper[4967]: I1008 00:07:31.117453 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 00:07:31 crc kubenswrapper[4967]: E1008 00:07:31.117620 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 00:07:31 crc kubenswrapper[4967]: E1008 00:07:31.117773 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 00:07:31 crc kubenswrapper[4967]: I1008 00:07:31.163493 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:31 crc kubenswrapper[4967]: I1008 00:07:31.163571 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:31 crc kubenswrapper[4967]: I1008 00:07:31.163583 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:31 crc kubenswrapper[4967]: I1008 00:07:31.163604 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:31 crc kubenswrapper[4967]: I1008 00:07:31.163622 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:31Z","lastTransitionTime":"2025-10-08T00:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:31 crc kubenswrapper[4967]: I1008 00:07:31.266409 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:31 crc kubenswrapper[4967]: I1008 00:07:31.266459 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:31 crc kubenswrapper[4967]: I1008 00:07:31.266474 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:31 crc kubenswrapper[4967]: I1008 00:07:31.266497 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:31 crc kubenswrapper[4967]: I1008 00:07:31.266512 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:31Z","lastTransitionTime":"2025-10-08T00:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:31 crc kubenswrapper[4967]: I1008 00:07:31.370075 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:31 crc kubenswrapper[4967]: I1008 00:07:31.370161 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:31 crc kubenswrapper[4967]: I1008 00:07:31.370186 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:31 crc kubenswrapper[4967]: I1008 00:07:31.370220 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:31 crc kubenswrapper[4967]: I1008 00:07:31.370243 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:31Z","lastTransitionTime":"2025-10-08T00:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:31 crc kubenswrapper[4967]: I1008 00:07:31.473025 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:31 crc kubenswrapper[4967]: I1008 00:07:31.473093 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:31 crc kubenswrapper[4967]: I1008 00:07:31.473112 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:31 crc kubenswrapper[4967]: I1008 00:07:31.473141 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:31 crc kubenswrapper[4967]: I1008 00:07:31.473164 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:31Z","lastTransitionTime":"2025-10-08T00:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:31 crc kubenswrapper[4967]: I1008 00:07:31.575818 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:31 crc kubenswrapper[4967]: I1008 00:07:31.575879 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:31 crc kubenswrapper[4967]: I1008 00:07:31.575902 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:31 crc kubenswrapper[4967]: I1008 00:07:31.575931 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:31 crc kubenswrapper[4967]: I1008 00:07:31.575952 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:31Z","lastTransitionTime":"2025-10-08T00:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:31 crc kubenswrapper[4967]: I1008 00:07:31.678358 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:31 crc kubenswrapper[4967]: I1008 00:07:31.678424 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:31 crc kubenswrapper[4967]: I1008 00:07:31.678443 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:31 crc kubenswrapper[4967]: I1008 00:07:31.678464 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:31 crc kubenswrapper[4967]: I1008 00:07:31.678480 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:31Z","lastTransitionTime":"2025-10-08T00:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:31 crc kubenswrapper[4967]: I1008 00:07:31.781438 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:31 crc kubenswrapper[4967]: I1008 00:07:31.781491 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:31 crc kubenswrapper[4967]: I1008 00:07:31.781507 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:31 crc kubenswrapper[4967]: I1008 00:07:31.781527 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:31 crc kubenswrapper[4967]: I1008 00:07:31.781543 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:31Z","lastTransitionTime":"2025-10-08T00:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:31 crc kubenswrapper[4967]: I1008 00:07:31.884556 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:31 crc kubenswrapper[4967]: I1008 00:07:31.884655 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:31 crc kubenswrapper[4967]: I1008 00:07:31.884723 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:31 crc kubenswrapper[4967]: I1008 00:07:31.884756 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:31 crc kubenswrapper[4967]: I1008 00:07:31.884780 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:31Z","lastTransitionTime":"2025-10-08T00:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:31 crc kubenswrapper[4967]: I1008 00:07:31.987521 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:31 crc kubenswrapper[4967]: I1008 00:07:31.987582 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:31 crc kubenswrapper[4967]: I1008 00:07:31.987601 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:31 crc kubenswrapper[4967]: I1008 00:07:31.987627 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:31 crc kubenswrapper[4967]: I1008 00:07:31.987647 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:31Z","lastTransitionTime":"2025-10-08T00:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:32 crc kubenswrapper[4967]: I1008 00:07:32.091127 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:32 crc kubenswrapper[4967]: I1008 00:07:32.091180 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:32 crc kubenswrapper[4967]: I1008 00:07:32.091203 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:32 crc kubenswrapper[4967]: I1008 00:07:32.091232 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:32 crc kubenswrapper[4967]: I1008 00:07:32.091254 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:32Z","lastTransitionTime":"2025-10-08T00:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:32 crc kubenswrapper[4967]: I1008 00:07:32.117478 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2h7z8" Oct 08 00:07:32 crc kubenswrapper[4967]: I1008 00:07:32.117552 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 00:07:32 crc kubenswrapper[4967]: E1008 00:07:32.118262 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 00:07:32 crc kubenswrapper[4967]: E1008 00:07:32.118084 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2h7z8" podUID="59f68771-e25d-4367-87ee-68a31d767882" Oct 08 00:07:32 crc kubenswrapper[4967]: I1008 00:07:32.139652 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fa4b747-ab6e-4a11-9903-ef2dcb9668c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbadcdade274b6d9be940f22bc949e3c55cadf95dcad43c8cb7a1085c0e6708c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d0a0b2ef4a98ee828688081a74f4b7ecc31c8fae415d7257a560ecd43960ce3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96b53b1b283025b3d078a3f505d469d1872ad4bb8ab8e612c36c70b0b00a920\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef86a2422616e1e0d5c3afe844818c2bdf2150f941d1aefc506995f58587b2bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1400e110997c64fa33d6b8672c7e91cab5f782952ddf8e1dea2a882689026e7f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 00:06:56.760088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 00:06:56.761210 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-369805205/tls.crt::/tmp/serving-cert-369805205/tls.key\\\\\\\"\\\\nI1008 00:07:02.321448 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 00:07:02.325932 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 00:07:02.325980 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 00:07:02.326025 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 00:07:02.326042 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 00:07:02.336203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 00:07:02.336240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 00:07:02.336245 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 00:07:02.336250 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 00:07:02.336253 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 00:07:02.336256 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 00:07:02.336258 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 00:07:02.336430 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 00:07:02.340774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:46Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca28d8db19036e47a7ef3dadb70941f6c6cf5735e7703977d036185c46fce40\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5b222ad84c31d8746dfe03124c7287b401deb3ca2c62ce2b4fbd40e638cee0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5b222ad84c31d8746dfe03124c7287b401deb3ca2c62ce2b4fbd40e638cee0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:32Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:32 crc kubenswrapper[4967]: I1008 00:07:32.157282 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:32Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:32 crc kubenswrapper[4967]: I1008 00:07:32.171113 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbfcdc882e2be1cb9cf666c9e7775671078f7dbbfe8421dd15db737b89920a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:32Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:32 crc kubenswrapper[4967]: I1008 00:07:32.184183 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9461162a-c27d-465e-9ca7-2d3936877cd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9778d78ddc7325abd4d77bc5e71b0b114f50be5dda4a438c7cb6b05640c8a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z9tp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28760af04ef321d0f91b2bce3ad074d44b63dabf676cc37f3be6aa8726341a2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z9tp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xvdtp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:32Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:32 crc kubenswrapper[4967]: I1008 00:07:32.193639 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:32 crc kubenswrapper[4967]: I1008 00:07:32.193681 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:32 crc kubenswrapper[4967]: I1008 00:07:32.193691 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:32 crc kubenswrapper[4967]: I1008 00:07:32.193705 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:32 crc kubenswrapper[4967]: I1008 00:07:32.193715 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:32Z","lastTransitionTime":"2025-10-08T00:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:32 crc kubenswrapper[4967]: I1008 00:07:32.201151 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94e63bc5641ce80d507f76a19b203f78c42368a7a4105f3fd09e19af21a1b735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:32Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:32 crc kubenswrapper[4967]: I1008 00:07:32.216758 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e98666fd3152bab57d3723e7da72aae2703b79f7ca73eb3c3e5f0e062a82592\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cf1cbe5548146fe359aafa351864125ec6b8bdd7b796aa4d311f6f3a16ad212\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:32Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:32 crc kubenswrapper[4967]: I1008 00:07:32.231992 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9tdbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41264a12-9486-4fee-a08c-607b590e5a70\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://806e85df796ae300d279f73a0cc4433a5bb1f7620ee8ce33c8dc4674c4ce0596\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5blv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9tdbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:32Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:32 crc kubenswrapper[4967]: I1008 00:07:32.283642 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f1af2ba-8907-4d14-afa1-ffde207c80f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4535cffb8c9d64e662e282ecb44f6817b4604866f0bb75015f3b888d5554f821\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b03b83ddff429dbfa8fe052901aaa86a39ff0dd75477ffa8a6b6f4198f5819ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c9fa0559eb35ba67e11e36069ac8d0b2fec76399788eef668a08ad4353df3ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3312263e12bd42fbed0476ab5ace66b5497f62c9477865e130d67e3922e05cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e69dc603055aa221d44d11c049ddcc50cec568979911a310ed773a90341a499\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115f7a33f3dedae8f59d2f45f465f95f48b99c824e895ef7a0be9fe7a1b61d23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1cfd26b02d0e17b111b8ad84dab6c2adf692de66ff77158ee7cbff941c66ecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1cfd26b02d0e17b111b8ad84dab6c2adf692de66ff77158ee7cbff941c66ecd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T00:07:29Z\\\",\\\"message\\\":\\\"iled objects of type *v1.Pod\\\\nI1008 00:07:28.973629 6638 obj_retry.go:409] Going to retry *v1.Pod resource setup for 1 objects: [openshift-multus/network-metrics-daemon-2h7z8]\\\\nI1008 00:07:28.973646 6638 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1008 00:07:28.973685 6638 obj_retry.go:285] Attempting retry of *v1.Pod openshift-multus/network-metrics-daemon-2h7z8 before timer (time: 2025-10-08 00:07:30.377116097 +0000 UTC m=+1.948919645): skip\\\\nI1008 00:07:28.973706 6638 obj_retry.go:420] Function iterateRetryResources for *v1.Pod ended (in 79.172µs)\\\\nI1008 00:07:28.973904 6638 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1008 00:07:28.973958 6638 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1008 00:07:28.974084 6638 factory.go:656] Stopping watch factory\\\\nI1008 00:07:28.974107 6638 ovnkube.go:599] Stopped ovnkube\\\\nI1008 00:07:28.974137 6638 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1008 00:07:28.974149 6638 handler.go:208] Removed *v1.Node event handler 2\\\\nI1008 00:07:28.974163 6638 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1008 00:07:28.974267 6638 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-bs8w9_openshift-ovn-kubernetes(4f1af2ba-8907-4d14-afa1-ffde207c80f4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27fb2b1225e386dcd7fecaaa0aa54ba1fdfafabfde6c3824ae07315313067542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bs8w9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:32Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:32 crc kubenswrapper[4967]: I1008 00:07:32.296469 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:32 crc kubenswrapper[4967]: I1008 00:07:32.296754 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:32 crc kubenswrapper[4967]: I1008 00:07:32.296859 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:32 crc kubenswrapper[4967]: I1008 00:07:32.296932 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:32 crc kubenswrapper[4967]: I1008 00:07:32.296996 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:32Z","lastTransitionTime":"2025-10-08T00:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:32 crc kubenswrapper[4967]: I1008 00:07:32.300043 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-v5n7t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abe7ae2d-7e63-4249-96e1-4000e9628ccc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52575e853ff6727dde7c89e4add4362be126127db27191c2a7bf2871efa719d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpsjd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:06Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-v5n7t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:32Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:32 crc kubenswrapper[4967]: I1008 00:07:32.313725 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:32Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:32 crc kubenswrapper[4967]: I1008 00:07:32.328572 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-npz2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c71501b1-aa9e-420e-b1bd-7e15365e7adc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dfe6c420d804b2d7daef4b5cfe3f87b77b369287c4d65553fa93f5f24f589d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gblk2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-npz2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:32Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:32 crc kubenswrapper[4967]: I1008 00:07:32.340520 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2h7z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59f68771-e25d-4367-87ee-68a31d767882\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7r2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7r2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:18Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2h7z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:32Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:32 crc kubenswrapper[4967]: I1008 00:07:32.351631 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58c4704-a88a-43b4-b944-4de626f13f7e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22f5dabb39da0a7bd01db4dae1437b9a231a1fcc0dfe749a4a5a637a72cf33c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01a2b3dc785520c79eae998482e5a1347725c76584357073238fe60a619edb77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6aa6f5e81feceab3d1c7587152a70546b365e629a69267e4d3084324a1de764a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63c2b7c4b2791bd1f16888c58b878dbff1f87884e68c9949de914d0d3e1d39fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:32Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:32 crc kubenswrapper[4967]: I1008 00:07:32.364561 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:32Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:32 crc kubenswrapper[4967]: I1008 00:07:32.379055 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9mrr7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9299bc4f-f8d3-4baf-9885-39573982fd64\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://994beafece9dead0d85ae74dc453a142460705fd66b52ba22f2fa799cf27da08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb49c8476b482416588c4d403016940730064a619c0641eb029c947a4bab1215\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb49c8476b482416588c4d403016940730064a619c0641eb029c947a4bab1215\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b7b63962dcbf477f1fa4d71826a4aa3358920a2c1b2f5e48716411b5bdca049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b7b63962dcbf477f1fa4d71826a4aa3358920a2c1b2f5e48716411b5bdca049\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1e770dc737173fd4f3220ed6c52b422ccbd829722386a37029e1c3e10d80e1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1e770dc737173fd4f3220ed6c52b422ccbd829722386a37029e1c3e10d80e1a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e357d970878c996f219a79f1876b363176c9847b9575f33090f62bc299529a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e357d970878c996f219a79f1876b363176c9847b9575f33090f62bc299529a95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f098966bfad49d06278eaa72be042596d488115e597737d9a0a0cfc78f99f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49f098966bfad49d06278eaa72be042596d488115e597737d9a0a0cfc78f99f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1d53876b7b863c612d597fd716c3260e8857b6542802073bb730f6756071b46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c1d53876b7b863c612d597fd716c3260e8857b6542802073bb730f6756071b46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9mrr7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:32Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:32 crc kubenswrapper[4967]: I1008 00:07:32.389800 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kpmlg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f975d47d-42a6-422e-80e2-d1ee282817fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a317ae6640eb38b3aee0f9738cca890e8ab3d92075b620cd312f40a0628d46c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qznqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ef61854aea40924c5372b0553cd0b7e8faa39a1d42fec5b65cf867c368f72c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qznqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-kpmlg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:32Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:32 crc kubenswrapper[4967]: I1008 00:07:32.399403 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:32 crc kubenswrapper[4967]: I1008 00:07:32.399439 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:32 crc kubenswrapper[4967]: I1008 00:07:32.399450 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:32 crc kubenswrapper[4967]: I1008 00:07:32.399468 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:32 crc kubenswrapper[4967]: I1008 00:07:32.399479 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:32Z","lastTransitionTime":"2025-10-08T00:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:32 crc kubenswrapper[4967]: I1008 00:07:32.502494 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:32 crc kubenswrapper[4967]: I1008 00:07:32.502543 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:32 crc kubenswrapper[4967]: I1008 00:07:32.502562 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:32 crc kubenswrapper[4967]: I1008 00:07:32.502584 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:32 crc kubenswrapper[4967]: I1008 00:07:32.502600 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:32Z","lastTransitionTime":"2025-10-08T00:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:32 crc kubenswrapper[4967]: I1008 00:07:32.606092 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:32 crc kubenswrapper[4967]: I1008 00:07:32.606212 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:32 crc kubenswrapper[4967]: I1008 00:07:32.606277 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:32 crc kubenswrapper[4967]: I1008 00:07:32.606314 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:32 crc kubenswrapper[4967]: I1008 00:07:32.606334 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:32Z","lastTransitionTime":"2025-10-08T00:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:32 crc kubenswrapper[4967]: I1008 00:07:32.708931 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:32 crc kubenswrapper[4967]: I1008 00:07:32.708973 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:32 crc kubenswrapper[4967]: I1008 00:07:32.708986 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:32 crc kubenswrapper[4967]: I1008 00:07:32.709004 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:32 crc kubenswrapper[4967]: I1008 00:07:32.709017 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:32Z","lastTransitionTime":"2025-10-08T00:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:32 crc kubenswrapper[4967]: I1008 00:07:32.812353 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:32 crc kubenswrapper[4967]: I1008 00:07:32.812414 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:32 crc kubenswrapper[4967]: I1008 00:07:32.812435 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:32 crc kubenswrapper[4967]: I1008 00:07:32.812459 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:32 crc kubenswrapper[4967]: I1008 00:07:32.812476 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:32Z","lastTransitionTime":"2025-10-08T00:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:32 crc kubenswrapper[4967]: I1008 00:07:32.915046 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:32 crc kubenswrapper[4967]: I1008 00:07:32.915147 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:32 crc kubenswrapper[4967]: I1008 00:07:32.915170 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:32 crc kubenswrapper[4967]: I1008 00:07:32.915194 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:32 crc kubenswrapper[4967]: I1008 00:07:32.915212 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:32Z","lastTransitionTime":"2025-10-08T00:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:33 crc kubenswrapper[4967]: I1008 00:07:33.017534 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:33 crc kubenswrapper[4967]: I1008 00:07:33.017573 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:33 crc kubenswrapper[4967]: I1008 00:07:33.017585 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:33 crc kubenswrapper[4967]: I1008 00:07:33.017600 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:33 crc kubenswrapper[4967]: I1008 00:07:33.017611 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:33Z","lastTransitionTime":"2025-10-08T00:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:33 crc kubenswrapper[4967]: I1008 00:07:33.117249 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 00:07:33 crc kubenswrapper[4967]: I1008 00:07:33.117375 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 00:07:33 crc kubenswrapper[4967]: E1008 00:07:33.117421 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 00:07:33 crc kubenswrapper[4967]: E1008 00:07:33.117585 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 00:07:33 crc kubenswrapper[4967]: I1008 00:07:33.120329 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:33 crc kubenswrapper[4967]: I1008 00:07:33.120453 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:33 crc kubenswrapper[4967]: I1008 00:07:33.120554 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:33 crc kubenswrapper[4967]: I1008 00:07:33.120660 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:33 crc kubenswrapper[4967]: I1008 00:07:33.120781 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:33Z","lastTransitionTime":"2025-10-08T00:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:33 crc kubenswrapper[4967]: I1008 00:07:33.223243 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:33 crc kubenswrapper[4967]: I1008 00:07:33.223313 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:33 crc kubenswrapper[4967]: I1008 00:07:33.223337 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:33 crc kubenswrapper[4967]: I1008 00:07:33.223370 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:33 crc kubenswrapper[4967]: I1008 00:07:33.223393 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:33Z","lastTransitionTime":"2025-10-08T00:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:33 crc kubenswrapper[4967]: I1008 00:07:33.326522 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:33 crc kubenswrapper[4967]: I1008 00:07:33.326571 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:33 crc kubenswrapper[4967]: I1008 00:07:33.326583 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:33 crc kubenswrapper[4967]: I1008 00:07:33.326599 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:33 crc kubenswrapper[4967]: I1008 00:07:33.326610 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:33Z","lastTransitionTime":"2025-10-08T00:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:33 crc kubenswrapper[4967]: I1008 00:07:33.429643 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:33 crc kubenswrapper[4967]: I1008 00:07:33.429724 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:33 crc kubenswrapper[4967]: I1008 00:07:33.429741 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:33 crc kubenswrapper[4967]: I1008 00:07:33.429761 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:33 crc kubenswrapper[4967]: I1008 00:07:33.429775 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:33Z","lastTransitionTime":"2025-10-08T00:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:33 crc kubenswrapper[4967]: I1008 00:07:33.532283 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:33 crc kubenswrapper[4967]: I1008 00:07:33.532344 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:33 crc kubenswrapper[4967]: I1008 00:07:33.532361 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:33 crc kubenswrapper[4967]: I1008 00:07:33.532387 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:33 crc kubenswrapper[4967]: I1008 00:07:33.532404 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:33Z","lastTransitionTime":"2025-10-08T00:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:33 crc kubenswrapper[4967]: I1008 00:07:33.635959 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:33 crc kubenswrapper[4967]: I1008 00:07:33.636001 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:33 crc kubenswrapper[4967]: I1008 00:07:33.636015 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:33 crc kubenswrapper[4967]: I1008 00:07:33.636033 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:33 crc kubenswrapper[4967]: I1008 00:07:33.636045 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:33Z","lastTransitionTime":"2025-10-08T00:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:33 crc kubenswrapper[4967]: I1008 00:07:33.739401 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:33 crc kubenswrapper[4967]: I1008 00:07:33.739458 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:33 crc kubenswrapper[4967]: I1008 00:07:33.739467 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:33 crc kubenswrapper[4967]: I1008 00:07:33.739483 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:33 crc kubenswrapper[4967]: I1008 00:07:33.739492 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:33Z","lastTransitionTime":"2025-10-08T00:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:33 crc kubenswrapper[4967]: I1008 00:07:33.842819 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:33 crc kubenswrapper[4967]: I1008 00:07:33.842876 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:33 crc kubenswrapper[4967]: I1008 00:07:33.842896 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:33 crc kubenswrapper[4967]: I1008 00:07:33.842921 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:33 crc kubenswrapper[4967]: I1008 00:07:33.842938 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:33Z","lastTransitionTime":"2025-10-08T00:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:33 crc kubenswrapper[4967]: I1008 00:07:33.946945 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:33 crc kubenswrapper[4967]: I1008 00:07:33.947011 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:33 crc kubenswrapper[4967]: I1008 00:07:33.947029 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:33 crc kubenswrapper[4967]: I1008 00:07:33.947054 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:33 crc kubenswrapper[4967]: I1008 00:07:33.947080 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:33Z","lastTransitionTime":"2025-10-08T00:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:34 crc kubenswrapper[4967]: I1008 00:07:34.050611 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:34 crc kubenswrapper[4967]: I1008 00:07:34.050715 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:34 crc kubenswrapper[4967]: I1008 00:07:34.050736 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:34 crc kubenswrapper[4967]: I1008 00:07:34.050764 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:34 crc kubenswrapper[4967]: I1008 00:07:34.050782 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:34Z","lastTransitionTime":"2025-10-08T00:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:34 crc kubenswrapper[4967]: I1008 00:07:34.117486 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2h7z8" Oct 08 00:07:34 crc kubenswrapper[4967]: I1008 00:07:34.117780 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 00:07:34 crc kubenswrapper[4967]: E1008 00:07:34.117977 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2h7z8" podUID="59f68771-e25d-4367-87ee-68a31d767882" Oct 08 00:07:34 crc kubenswrapper[4967]: E1008 00:07:34.118168 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 00:07:34 crc kubenswrapper[4967]: I1008 00:07:34.154808 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:34 crc kubenswrapper[4967]: I1008 00:07:34.154901 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:34 crc kubenswrapper[4967]: I1008 00:07:34.154928 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:34 crc kubenswrapper[4967]: I1008 00:07:34.154963 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:34 crc kubenswrapper[4967]: I1008 00:07:34.154987 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:34Z","lastTransitionTime":"2025-10-08T00:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:34 crc kubenswrapper[4967]: I1008 00:07:34.257372 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:34 crc kubenswrapper[4967]: I1008 00:07:34.257450 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:34 crc kubenswrapper[4967]: I1008 00:07:34.257463 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:34 crc kubenswrapper[4967]: I1008 00:07:34.257479 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:34 crc kubenswrapper[4967]: I1008 00:07:34.257491 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:34Z","lastTransitionTime":"2025-10-08T00:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:34 crc kubenswrapper[4967]: I1008 00:07:34.360322 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:34 crc kubenswrapper[4967]: I1008 00:07:34.360350 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:34 crc kubenswrapper[4967]: I1008 00:07:34.360359 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:34 crc kubenswrapper[4967]: I1008 00:07:34.360372 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:34 crc kubenswrapper[4967]: I1008 00:07:34.360382 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:34Z","lastTransitionTime":"2025-10-08T00:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:34 crc kubenswrapper[4967]: I1008 00:07:34.463270 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:34 crc kubenswrapper[4967]: I1008 00:07:34.463350 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:34 crc kubenswrapper[4967]: I1008 00:07:34.463364 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:34 crc kubenswrapper[4967]: I1008 00:07:34.463380 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:34 crc kubenswrapper[4967]: I1008 00:07:34.463394 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:34Z","lastTransitionTime":"2025-10-08T00:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:34 crc kubenswrapper[4967]: I1008 00:07:34.566348 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:34 crc kubenswrapper[4967]: I1008 00:07:34.566426 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:34 crc kubenswrapper[4967]: I1008 00:07:34.566440 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:34 crc kubenswrapper[4967]: I1008 00:07:34.566458 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:34 crc kubenswrapper[4967]: I1008 00:07:34.566471 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:34Z","lastTransitionTime":"2025-10-08T00:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:34 crc kubenswrapper[4967]: I1008 00:07:34.669097 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:34 crc kubenswrapper[4967]: I1008 00:07:34.669147 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:34 crc kubenswrapper[4967]: I1008 00:07:34.669167 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:34 crc kubenswrapper[4967]: I1008 00:07:34.669185 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:34 crc kubenswrapper[4967]: I1008 00:07:34.669198 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:34Z","lastTransitionTime":"2025-10-08T00:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:34 crc kubenswrapper[4967]: I1008 00:07:34.734297 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 08 00:07:34 crc kubenswrapper[4967]: I1008 00:07:34.743832 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Oct 08 00:07:34 crc kubenswrapper[4967]: I1008 00:07:34.747167 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94e63bc5641ce80d507f76a19b203f78c42368a7a4105f3fd09e19af21a1b735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:34Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:34 crc kubenswrapper[4967]: I1008 00:07:34.765939 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e98666fd3152bab57d3723e7da72aae2703b79f7ca73eb3c3e5f0e062a82592\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cf1cbe5548146fe359aafa351864125ec6b8bdd7b796aa4d311f6f3a16ad212\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:34Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:34 crc kubenswrapper[4967]: I1008 00:07:34.771290 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:34 crc kubenswrapper[4967]: I1008 00:07:34.771337 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:34 crc kubenswrapper[4967]: I1008 00:07:34.771349 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:34 crc kubenswrapper[4967]: I1008 00:07:34.771365 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:34 crc kubenswrapper[4967]: I1008 00:07:34.771376 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:34Z","lastTransitionTime":"2025-10-08T00:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:34 crc kubenswrapper[4967]: I1008 00:07:34.782162 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9tdbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41264a12-9486-4fee-a08c-607b590e5a70\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://806e85df796ae300d279f73a0cc4433a5bb1f7620ee8ce33c8dc4674c4ce0596\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5blv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9tdbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:34Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:34 crc kubenswrapper[4967]: I1008 00:07:34.802054 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f1af2ba-8907-4d14-afa1-ffde207c80f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4535cffb8c9d64e662e282ecb44f6817b4604866f0bb75015f3b888d5554f821\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b03b83ddff429dbfa8fe052901aaa86a39ff0dd75477ffa8a6b6f4198f5819ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c9fa0559eb35ba67e11e36069ac8d0b2fec76399788eef668a08ad4353df3ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3312263e12bd42fbed0476ab5ace66b5497f62c9477865e130d67e3922e05cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e69dc603055aa221d44d11c049ddcc50cec568979911a310ed773a90341a499\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115f7a33f3dedae8f59d2f45f465f95f48b99c824e895ef7a0be9fe7a1b61d23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1cfd26b02d0e17b111b8ad84dab6c2adf692de66ff77158ee7cbff941c66ecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1cfd26b02d0e17b111b8ad84dab6c2adf692de66ff77158ee7cbff941c66ecd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T00:07:29Z\\\",\\\"message\\\":\\\"iled objects of type *v1.Pod\\\\nI1008 00:07:28.973629 6638 obj_retry.go:409] Going to retry *v1.Pod resource setup for 1 objects: [openshift-multus/network-metrics-daemon-2h7z8]\\\\nI1008 00:07:28.973646 6638 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1008 00:07:28.973685 6638 obj_retry.go:285] Attempting retry of *v1.Pod openshift-multus/network-metrics-daemon-2h7z8 before timer (time: 2025-10-08 00:07:30.377116097 +0000 UTC m=+1.948919645): skip\\\\nI1008 00:07:28.973706 6638 obj_retry.go:420] Function iterateRetryResources for *v1.Pod ended (in 79.172µs)\\\\nI1008 00:07:28.973904 6638 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1008 00:07:28.973958 6638 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1008 00:07:28.974084 6638 factory.go:656] Stopping watch factory\\\\nI1008 00:07:28.974107 6638 ovnkube.go:599] Stopped ovnkube\\\\nI1008 00:07:28.974137 6638 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1008 00:07:28.974149 6638 handler.go:208] Removed *v1.Node event handler 2\\\\nI1008 00:07:28.974163 6638 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1008 00:07:28.974267 6638 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-bs8w9_openshift-ovn-kubernetes(4f1af2ba-8907-4d14-afa1-ffde207c80f4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27fb2b1225e386dcd7fecaaa0aa54ba1fdfafabfde6c3824ae07315313067542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bs8w9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:34Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:34 crc kubenswrapper[4967]: I1008 00:07:34.812503 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-v5n7t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abe7ae2d-7e63-4249-96e1-4000e9628ccc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52575e853ff6727dde7c89e4add4362be126127db27191c2a7bf2871efa719d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpsjd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:06Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-v5n7t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:34Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:34 crc kubenswrapper[4967]: I1008 00:07:34.824943 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:34Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:34 crc kubenswrapper[4967]: I1008 00:07:34.835067 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/59f68771-e25d-4367-87ee-68a31d767882-metrics-certs\") pod \"network-metrics-daemon-2h7z8\" (UID: \"59f68771-e25d-4367-87ee-68a31d767882\") " pod="openshift-multus/network-metrics-daemon-2h7z8" Oct 08 00:07:34 crc kubenswrapper[4967]: E1008 00:07:34.835219 4967 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 00:07:34 crc kubenswrapper[4967]: E1008 00:07:34.835285 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/59f68771-e25d-4367-87ee-68a31d767882-metrics-certs podName:59f68771-e25d-4367-87ee-68a31d767882 nodeName:}" failed. No retries permitted until 2025-10-08 00:07:50.835268167 +0000 UTC m=+69.485802816 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/59f68771-e25d-4367-87ee-68a31d767882-metrics-certs") pod "network-metrics-daemon-2h7z8" (UID: "59f68771-e25d-4367-87ee-68a31d767882") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 00:07:34 crc kubenswrapper[4967]: I1008 00:07:34.840882 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-npz2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c71501b1-aa9e-420e-b1bd-7e15365e7adc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dfe6c420d804b2d7daef4b5cfe3f87b77b369287c4d65553fa93f5f24f589d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gblk2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-npz2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:34Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:34 crc kubenswrapper[4967]: I1008 00:07:34.858512 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2h7z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59f68771-e25d-4367-87ee-68a31d767882\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7r2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7r2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:18Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2h7z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:34Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:34 crc kubenswrapper[4967]: I1008 00:07:34.872389 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58c4704-a88a-43b4-b944-4de626f13f7e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22f5dabb39da0a7bd01db4dae1437b9a231a1fcc0dfe749a4a5a637a72cf33c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01a2b3dc785520c79eae998482e5a1347725c76584357073238fe60a619edb77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6aa6f5e81feceab3d1c7587152a70546b365e629a69267e4d3084324a1de764a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63c2b7c4b2791bd1f16888c58b878dbff1f87884e68c9949de914d0d3e1d39fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:34Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:34 crc kubenswrapper[4967]: I1008 00:07:34.874172 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:34 crc kubenswrapper[4967]: I1008 00:07:34.874213 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:34 crc kubenswrapper[4967]: I1008 00:07:34.874264 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:34 crc kubenswrapper[4967]: I1008 00:07:34.874285 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:34 crc kubenswrapper[4967]: I1008 00:07:34.874298 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:34Z","lastTransitionTime":"2025-10-08T00:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:34 crc kubenswrapper[4967]: I1008 00:07:34.887136 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:34Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:34 crc kubenswrapper[4967]: I1008 00:07:34.906626 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9mrr7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9299bc4f-f8d3-4baf-9885-39573982fd64\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://994beafece9dead0d85ae74dc453a142460705fd66b52ba22f2fa799cf27da08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb49c8476b482416588c4d403016940730064a619c0641eb029c947a4bab1215\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb49c8476b482416588c4d403016940730064a619c0641eb029c947a4bab1215\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b7b63962dcbf477f1fa4d71826a4aa3358920a2c1b2f5e48716411b5bdca049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b7b63962dcbf477f1fa4d71826a4aa3358920a2c1b2f5e48716411b5bdca049\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1e770dc737173fd4f3220ed6c52b422ccbd829722386a37029e1c3e10d80e1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1e770dc737173fd4f3220ed6c52b422ccbd829722386a37029e1c3e10d80e1a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e357d970878c996f219a79f1876b363176c9847b9575f33090f62bc299529a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e357d970878c996f219a79f1876b363176c9847b9575f33090f62bc299529a95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f098966bfad49d06278eaa72be042596d488115e597737d9a0a0cfc78f99f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49f098966bfad49d06278eaa72be042596d488115e597737d9a0a0cfc78f99f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1d53876b7b863c612d597fd716c3260e8857b6542802073bb730f6756071b46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c1d53876b7b863c612d597fd716c3260e8857b6542802073bb730f6756071b46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9mrr7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:34Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:34 crc kubenswrapper[4967]: I1008 00:07:34.921924 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kpmlg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f975d47d-42a6-422e-80e2-d1ee282817fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a317ae6640eb38b3aee0f9738cca890e8ab3d92075b620cd312f40a0628d46c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qznqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ef61854aea40924c5372b0553cd0b7e8faa39a1d42fec5b65cf867c368f72c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qznqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-kpmlg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:34Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:34 crc kubenswrapper[4967]: I1008 00:07:34.938509 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fa4b747-ab6e-4a11-9903-ef2dcb9668c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbadcdade274b6d9be940f22bc949e3c55cadf95dcad43c8cb7a1085c0e6708c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d0a0b2ef4a98ee828688081a74f4b7ecc31c8fae415d7257a560ecd43960ce3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96b53b1b283025b3d078a3f505d469d1872ad4bb8ab8e612c36c70b0b00a920\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef86a2422616e1e0d5c3afe844818c2bdf2150f941d1aefc506995f58587b2bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1400e110997c64fa33d6b8672c7e91cab5f782952ddf8e1dea2a882689026e7f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 00:06:56.760088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 00:06:56.761210 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-369805205/tls.crt::/tmp/serving-cert-369805205/tls.key\\\\\\\"\\\\nI1008 00:07:02.321448 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 00:07:02.325932 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 00:07:02.325980 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 00:07:02.326025 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 00:07:02.326042 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 00:07:02.336203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 00:07:02.336240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 00:07:02.336245 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 00:07:02.336250 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 00:07:02.336253 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 00:07:02.336256 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 00:07:02.336258 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 00:07:02.336430 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 00:07:02.340774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:46Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca28d8db19036e47a7ef3dadb70941f6c6cf5735e7703977d036185c46fce40\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5b222ad84c31d8746dfe03124c7287b401deb3ca2c62ce2b4fbd40e638cee0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5b222ad84c31d8746dfe03124c7287b401deb3ca2c62ce2b4fbd40e638cee0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:34Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:34 crc kubenswrapper[4967]: I1008 00:07:34.955284 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:34Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:34 crc kubenswrapper[4967]: I1008 00:07:34.975089 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbfcdc882e2be1cb9cf666c9e7775671078f7dbbfe8421dd15db737b89920a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:34Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:34 crc kubenswrapper[4967]: I1008 00:07:34.976428 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:34 crc kubenswrapper[4967]: I1008 00:07:34.976457 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:34 crc kubenswrapper[4967]: I1008 00:07:34.976468 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:34 crc kubenswrapper[4967]: I1008 00:07:34.976481 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:34 crc kubenswrapper[4967]: I1008 00:07:34.976490 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:34Z","lastTransitionTime":"2025-10-08T00:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:34 crc kubenswrapper[4967]: I1008 00:07:34.988277 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9461162a-c27d-465e-9ca7-2d3936877cd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9778d78ddc7325abd4d77bc5e71b0b114f50be5dda4a438c7cb6b05640c8a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z9tp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28760af04ef321d0f91b2bce3ad074d44b63dabf676cc37f3be6aa8726341a2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z9tp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xvdtp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:34Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:35 crc kubenswrapper[4967]: I1008 00:07:35.036831 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 00:07:35 crc kubenswrapper[4967]: E1008 00:07:35.036996 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 00:08:07.03696697 +0000 UTC m=+85.687501629 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:07:35 crc kubenswrapper[4967]: I1008 00:07:35.037300 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 00:07:35 crc kubenswrapper[4967]: I1008 00:07:35.037372 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 00:07:35 crc kubenswrapper[4967]: I1008 00:07:35.037404 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 00:07:35 crc kubenswrapper[4967]: I1008 00:07:35.037436 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 00:07:35 crc kubenswrapper[4967]: E1008 00:07:35.037460 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 00:07:35 crc kubenswrapper[4967]: E1008 00:07:35.037497 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 00:07:35 crc kubenswrapper[4967]: E1008 00:07:35.037512 4967 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 00:07:35 crc kubenswrapper[4967]: E1008 00:07:35.037525 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 00:07:35 crc kubenswrapper[4967]: E1008 00:07:35.037540 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 00:07:35 crc kubenswrapper[4967]: E1008 00:07:35.037550 4967 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 00:07:35 crc kubenswrapper[4967]: E1008 00:07:35.037661 4967 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 00:07:35 crc kubenswrapper[4967]: E1008 00:07:35.037743 4967 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 00:07:35 crc kubenswrapper[4967]: E1008 00:07:35.037605 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-08 00:08:07.037544575 +0000 UTC m=+85.688079224 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 00:07:35 crc kubenswrapper[4967]: E1008 00:07:35.037773 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-08 00:08:07.037764931 +0000 UTC m=+85.688299580 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 00:07:35 crc kubenswrapper[4967]: E1008 00:07:35.037787 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 00:08:07.037779481 +0000 UTC m=+85.688314210 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 00:07:35 crc kubenswrapper[4967]: E1008 00:07:35.037799 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 00:08:07.037792752 +0000 UTC m=+85.688327411 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 00:07:35 crc kubenswrapper[4967]: I1008 00:07:35.079158 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:35 crc kubenswrapper[4967]: I1008 00:07:35.079218 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:35 crc kubenswrapper[4967]: I1008 00:07:35.079232 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:35 crc kubenswrapper[4967]: I1008 00:07:35.079249 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:35 crc kubenswrapper[4967]: I1008 00:07:35.079531 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:35Z","lastTransitionTime":"2025-10-08T00:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:35 crc kubenswrapper[4967]: I1008 00:07:35.117463 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 00:07:35 crc kubenswrapper[4967]: E1008 00:07:35.117611 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 00:07:35 crc kubenswrapper[4967]: I1008 00:07:35.117865 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 00:07:35 crc kubenswrapper[4967]: E1008 00:07:35.118035 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 00:07:35 crc kubenswrapper[4967]: I1008 00:07:35.182465 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:35 crc kubenswrapper[4967]: I1008 00:07:35.182505 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:35 crc kubenswrapper[4967]: I1008 00:07:35.182515 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:35 crc kubenswrapper[4967]: I1008 00:07:35.182530 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:35 crc kubenswrapper[4967]: I1008 00:07:35.182541 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:35Z","lastTransitionTime":"2025-10-08T00:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:35 crc kubenswrapper[4967]: I1008 00:07:35.285475 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:35 crc kubenswrapper[4967]: I1008 00:07:35.285548 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:35 crc kubenswrapper[4967]: I1008 00:07:35.285571 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:35 crc kubenswrapper[4967]: I1008 00:07:35.285601 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:35 crc kubenswrapper[4967]: I1008 00:07:35.285623 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:35Z","lastTransitionTime":"2025-10-08T00:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:35 crc kubenswrapper[4967]: I1008 00:07:35.388074 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:35 crc kubenswrapper[4967]: I1008 00:07:35.388153 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:35 crc kubenswrapper[4967]: I1008 00:07:35.388180 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:35 crc kubenswrapper[4967]: I1008 00:07:35.388211 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:35 crc kubenswrapper[4967]: I1008 00:07:35.388235 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:35Z","lastTransitionTime":"2025-10-08T00:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:35 crc kubenswrapper[4967]: I1008 00:07:35.491114 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:35 crc kubenswrapper[4967]: I1008 00:07:35.491165 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:35 crc kubenswrapper[4967]: I1008 00:07:35.491186 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:35 crc kubenswrapper[4967]: I1008 00:07:35.491211 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:35 crc kubenswrapper[4967]: I1008 00:07:35.491228 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:35Z","lastTransitionTime":"2025-10-08T00:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:35 crc kubenswrapper[4967]: I1008 00:07:35.593275 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:35 crc kubenswrapper[4967]: I1008 00:07:35.593319 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:35 crc kubenswrapper[4967]: I1008 00:07:35.593331 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:35 crc kubenswrapper[4967]: I1008 00:07:35.593347 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:35 crc kubenswrapper[4967]: I1008 00:07:35.593359 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:35Z","lastTransitionTime":"2025-10-08T00:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:35 crc kubenswrapper[4967]: I1008 00:07:35.695367 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:35 crc kubenswrapper[4967]: I1008 00:07:35.695513 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:35 crc kubenswrapper[4967]: I1008 00:07:35.695522 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:35 crc kubenswrapper[4967]: I1008 00:07:35.695535 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:35 crc kubenswrapper[4967]: I1008 00:07:35.695544 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:35Z","lastTransitionTime":"2025-10-08T00:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:35 crc kubenswrapper[4967]: I1008 00:07:35.798403 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:35 crc kubenswrapper[4967]: I1008 00:07:35.798438 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:35 crc kubenswrapper[4967]: I1008 00:07:35.798450 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:35 crc kubenswrapper[4967]: I1008 00:07:35.798465 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:35 crc kubenswrapper[4967]: I1008 00:07:35.798474 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:35Z","lastTransitionTime":"2025-10-08T00:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:35 crc kubenswrapper[4967]: I1008 00:07:35.901518 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:35 crc kubenswrapper[4967]: I1008 00:07:35.901578 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:35 crc kubenswrapper[4967]: I1008 00:07:35.901596 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:35 crc kubenswrapper[4967]: I1008 00:07:35.901619 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:35 crc kubenswrapper[4967]: I1008 00:07:35.901633 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:35Z","lastTransitionTime":"2025-10-08T00:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:36 crc kubenswrapper[4967]: I1008 00:07:36.005088 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:36 crc kubenswrapper[4967]: I1008 00:07:36.005157 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:36 crc kubenswrapper[4967]: I1008 00:07:36.005179 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:36 crc kubenswrapper[4967]: I1008 00:07:36.005211 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:36 crc kubenswrapper[4967]: I1008 00:07:36.005229 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:36Z","lastTransitionTime":"2025-10-08T00:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:36 crc kubenswrapper[4967]: I1008 00:07:36.108540 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:36 crc kubenswrapper[4967]: I1008 00:07:36.108596 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:36 crc kubenswrapper[4967]: I1008 00:07:36.108613 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:36 crc kubenswrapper[4967]: I1008 00:07:36.108630 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:36 crc kubenswrapper[4967]: I1008 00:07:36.108643 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:36Z","lastTransitionTime":"2025-10-08T00:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:36 crc kubenswrapper[4967]: I1008 00:07:36.116965 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2h7z8" Oct 08 00:07:36 crc kubenswrapper[4967]: I1008 00:07:36.116970 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 00:07:36 crc kubenswrapper[4967]: E1008 00:07:36.117203 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2h7z8" podUID="59f68771-e25d-4367-87ee-68a31d767882" Oct 08 00:07:36 crc kubenswrapper[4967]: E1008 00:07:36.117320 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 00:07:36 crc kubenswrapper[4967]: I1008 00:07:36.211641 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:36 crc kubenswrapper[4967]: I1008 00:07:36.211706 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:36 crc kubenswrapper[4967]: I1008 00:07:36.211719 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:36 crc kubenswrapper[4967]: I1008 00:07:36.211736 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:36 crc kubenswrapper[4967]: I1008 00:07:36.211747 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:36Z","lastTransitionTime":"2025-10-08T00:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:36 crc kubenswrapper[4967]: I1008 00:07:36.314122 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:36 crc kubenswrapper[4967]: I1008 00:07:36.314187 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:36 crc kubenswrapper[4967]: I1008 00:07:36.314205 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:36 crc kubenswrapper[4967]: I1008 00:07:36.314234 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:36 crc kubenswrapper[4967]: I1008 00:07:36.314253 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:36Z","lastTransitionTime":"2025-10-08T00:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:36 crc kubenswrapper[4967]: I1008 00:07:36.416104 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:36 crc kubenswrapper[4967]: I1008 00:07:36.416178 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:36 crc kubenswrapper[4967]: I1008 00:07:36.416199 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:36 crc kubenswrapper[4967]: I1008 00:07:36.416222 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:36 crc kubenswrapper[4967]: I1008 00:07:36.416239 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:36Z","lastTransitionTime":"2025-10-08T00:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:36 crc kubenswrapper[4967]: I1008 00:07:36.518496 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:36 crc kubenswrapper[4967]: I1008 00:07:36.518554 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:36 crc kubenswrapper[4967]: I1008 00:07:36.518568 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:36 crc kubenswrapper[4967]: I1008 00:07:36.518586 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:36 crc kubenswrapper[4967]: I1008 00:07:36.518599 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:36Z","lastTransitionTime":"2025-10-08T00:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:36 crc kubenswrapper[4967]: I1008 00:07:36.621570 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:36 crc kubenswrapper[4967]: I1008 00:07:36.621613 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:36 crc kubenswrapper[4967]: I1008 00:07:36.621623 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:36 crc kubenswrapper[4967]: I1008 00:07:36.621639 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:36 crc kubenswrapper[4967]: I1008 00:07:36.621651 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:36Z","lastTransitionTime":"2025-10-08T00:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:36 crc kubenswrapper[4967]: I1008 00:07:36.724411 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:36 crc kubenswrapper[4967]: I1008 00:07:36.724467 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:36 crc kubenswrapper[4967]: I1008 00:07:36.724483 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:36 crc kubenswrapper[4967]: I1008 00:07:36.724504 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:36 crc kubenswrapper[4967]: I1008 00:07:36.724522 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:36Z","lastTransitionTime":"2025-10-08T00:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:36 crc kubenswrapper[4967]: I1008 00:07:36.826601 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:36 crc kubenswrapper[4967]: I1008 00:07:36.826646 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:36 crc kubenswrapper[4967]: I1008 00:07:36.826655 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:36 crc kubenswrapper[4967]: I1008 00:07:36.826691 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:36 crc kubenswrapper[4967]: I1008 00:07:36.826700 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:36Z","lastTransitionTime":"2025-10-08T00:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:36 crc kubenswrapper[4967]: I1008 00:07:36.928952 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:36 crc kubenswrapper[4967]: I1008 00:07:36.929017 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:36 crc kubenswrapper[4967]: I1008 00:07:36.929039 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:36 crc kubenswrapper[4967]: I1008 00:07:36.929064 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:36 crc kubenswrapper[4967]: I1008 00:07:36.929083 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:36Z","lastTransitionTime":"2025-10-08T00:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:36 crc kubenswrapper[4967]: I1008 00:07:36.937894 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 00:07:36 crc kubenswrapper[4967]: I1008 00:07:36.951125 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:36Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:36 crc kubenswrapper[4967]: I1008 00:07:36.960614 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbfcdc882e2be1cb9cf666c9e7775671078f7dbbfe8421dd15db737b89920a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:36Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:36 crc kubenswrapper[4967]: I1008 00:07:36.971224 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9461162a-c27d-465e-9ca7-2d3936877cd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9778d78ddc7325abd4d77bc5e71b0b114f50be5dda4a438c7cb6b05640c8a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z9tp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28760af04ef321d0f91b2bce3ad074d44b63dabf676cc37f3be6aa8726341a2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z9tp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xvdtp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:36Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:36 crc kubenswrapper[4967]: I1008 00:07:36.986774 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fa4b747-ab6e-4a11-9903-ef2dcb9668c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbadcdade274b6d9be940f22bc949e3c55cadf95dcad43c8cb7a1085c0e6708c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d0a0b2ef4a98ee828688081a74f4b7ecc31c8fae415d7257a560ecd43960ce3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96b53b1b283025b3d078a3f505d469d1872ad4bb8ab8e612c36c70b0b00a920\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef86a2422616e1e0d5c3afe844818c2bdf2150f941d1aefc506995f58587b2bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1400e110997c64fa33d6b8672c7e91cab5f782952ddf8e1dea2a882689026e7f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 00:06:56.760088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 00:06:56.761210 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-369805205/tls.crt::/tmp/serving-cert-369805205/tls.key\\\\\\\"\\\\nI1008 00:07:02.321448 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 00:07:02.325932 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 00:07:02.325980 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 00:07:02.326025 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 00:07:02.326042 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 00:07:02.336203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 00:07:02.336240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 00:07:02.336245 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 00:07:02.336250 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 00:07:02.336253 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 00:07:02.336256 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 00:07:02.336258 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 00:07:02.336430 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 00:07:02.340774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:46Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca28d8db19036e47a7ef3dadb70941f6c6cf5735e7703977d036185c46fce40\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5b222ad84c31d8746dfe03124c7287b401deb3ca2c62ce2b4fbd40e638cee0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5b222ad84c31d8746dfe03124c7287b401deb3ca2c62ce2b4fbd40e638cee0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:36Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:36 crc kubenswrapper[4967]: I1008 00:07:36.995267 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9tdbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41264a12-9486-4fee-a08c-607b590e5a70\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://806e85df796ae300d279f73a0cc4433a5bb1f7620ee8ce33c8dc4674c4ce0596\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5blv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9tdbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:36Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:37 crc kubenswrapper[4967]: I1008 00:07:37.012078 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f1af2ba-8907-4d14-afa1-ffde207c80f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4535cffb8c9d64e662e282ecb44f6817b4604866f0bb75015f3b888d5554f821\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b03b83ddff429dbfa8fe052901aaa86a39ff0dd75477ffa8a6b6f4198f5819ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c9fa0559eb35ba67e11e36069ac8d0b2fec76399788eef668a08ad4353df3ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3312263e12bd42fbed0476ab5ace66b5497f62c9477865e130d67e3922e05cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e69dc603055aa221d44d11c049ddcc50cec568979911a310ed773a90341a499\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115f7a33f3dedae8f59d2f45f465f95f48b99c824e895ef7a0be9fe7a1b61d23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1cfd26b02d0e17b111b8ad84dab6c2adf692de66ff77158ee7cbff941c66ecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1cfd26b02d0e17b111b8ad84dab6c2adf692de66ff77158ee7cbff941c66ecd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T00:07:29Z\\\",\\\"message\\\":\\\"iled objects of type *v1.Pod\\\\nI1008 00:07:28.973629 6638 obj_retry.go:409] Going to retry *v1.Pod resource setup for 1 objects: [openshift-multus/network-metrics-daemon-2h7z8]\\\\nI1008 00:07:28.973646 6638 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1008 00:07:28.973685 6638 obj_retry.go:285] Attempting retry of *v1.Pod openshift-multus/network-metrics-daemon-2h7z8 before timer (time: 2025-10-08 00:07:30.377116097 +0000 UTC m=+1.948919645): skip\\\\nI1008 00:07:28.973706 6638 obj_retry.go:420] Function iterateRetryResources for *v1.Pod ended (in 79.172µs)\\\\nI1008 00:07:28.973904 6638 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1008 00:07:28.973958 6638 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1008 00:07:28.974084 6638 factory.go:656] Stopping watch factory\\\\nI1008 00:07:28.974107 6638 ovnkube.go:599] Stopped ovnkube\\\\nI1008 00:07:28.974137 6638 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1008 00:07:28.974149 6638 handler.go:208] Removed *v1.Node event handler 2\\\\nI1008 00:07:28.974163 6638 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1008 00:07:28.974267 6638 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-bs8w9_openshift-ovn-kubernetes(4f1af2ba-8907-4d14-afa1-ffde207c80f4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27fb2b1225e386dcd7fecaaa0aa54ba1fdfafabfde6c3824ae07315313067542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bs8w9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:37Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:37 crc kubenswrapper[4967]: I1008 00:07:37.021648 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-v5n7t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abe7ae2d-7e63-4249-96e1-4000e9628ccc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52575e853ff6727dde7c89e4add4362be126127db27191c2a7bf2871efa719d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpsjd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:06Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-v5n7t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:37Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:37 crc kubenswrapper[4967]: I1008 00:07:37.031255 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:37 crc kubenswrapper[4967]: I1008 00:07:37.031358 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:37 crc kubenswrapper[4967]: I1008 00:07:37.031381 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:37 crc kubenswrapper[4967]: I1008 00:07:37.031410 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:37 crc kubenswrapper[4967]: I1008 00:07:37.031428 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:37Z","lastTransitionTime":"2025-10-08T00:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:37 crc kubenswrapper[4967]: I1008 00:07:37.033462 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94e63bc5641ce80d507f76a19b203f78c42368a7a4105f3fd09e19af21a1b735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:37Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:37 crc kubenswrapper[4967]: I1008 00:07:37.045389 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e98666fd3152bab57d3723e7da72aae2703b79f7ca73eb3c3e5f0e062a82592\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cf1cbe5548146fe359aafa351864125ec6b8bdd7b796aa4d311f6f3a16ad212\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:37Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:37 crc kubenswrapper[4967]: I1008 00:07:37.060315 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-npz2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c71501b1-aa9e-420e-b1bd-7e15365e7adc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dfe6c420d804b2d7daef4b5cfe3f87b77b369287c4d65553fa93f5f24f589d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gblk2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-npz2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:37Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:37 crc kubenswrapper[4967]: I1008 00:07:37.071991 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2h7z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59f68771-e25d-4367-87ee-68a31d767882\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7r2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7r2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:18Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2h7z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:37Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:37 crc kubenswrapper[4967]: I1008 00:07:37.086426 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6ea2440-b2ae-481e-9442-0b3dda9688ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37fd5dcdfb864320d4d17729ab48c6da505fd867f384ad70a6dcf5bed78a3978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c7893dfaa0370f159d41c30cd869bb18550a819e8c3d20dd6f57dcbcc697a57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c430979d298477b4394c3567bb3800ab7dea874cee8fa41a6a50a19a7e3a0133\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://426e86ab409776ff0fa9505dc0d8ec36f27a5bf21d3aedafb19d3935bdba7529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://426e86ab409776ff0fa9505dc0d8ec36f27a5bf21d3aedafb19d3935bdba7529\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:37Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:37 crc kubenswrapper[4967]: I1008 00:07:37.101358 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:37Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:37 crc kubenswrapper[4967]: I1008 00:07:37.112573 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:37Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:37 crc kubenswrapper[4967]: I1008 00:07:37.117757 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 00:07:37 crc kubenswrapper[4967]: I1008 00:07:37.117768 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 00:07:37 crc kubenswrapper[4967]: E1008 00:07:37.117851 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 00:07:37 crc kubenswrapper[4967]: E1008 00:07:37.117935 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 00:07:37 crc kubenswrapper[4967]: I1008 00:07:37.126191 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9mrr7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9299bc4f-f8d3-4baf-9885-39573982fd64\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://994beafece9dead0d85ae74dc453a142460705fd66b52ba22f2fa799cf27da08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb49c8476b482416588c4d403016940730064a619c0641eb029c947a4bab1215\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb49c8476b482416588c4d403016940730064a619c0641eb029c947a4bab1215\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b7b63962dcbf477f1fa4d71826a4aa3358920a2c1b2f5e48716411b5bdca049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b7b63962dcbf477f1fa4d71826a4aa3358920a2c1b2f5e48716411b5bdca049\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1e770dc737173fd4f3220ed6c52b422ccbd829722386a37029e1c3e10d80e1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1e770dc737173fd4f3220ed6c52b422ccbd829722386a37029e1c3e10d80e1a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e357d970878c996f219a79f1876b363176c9847b9575f33090f62bc299529a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e357d970878c996f219a79f1876b363176c9847b9575f33090f62bc299529a95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f098966bfad49d06278eaa72be042596d488115e597737d9a0a0cfc78f99f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49f098966bfad49d06278eaa72be042596d488115e597737d9a0a0cfc78f99f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1d53876b7b863c612d597fd716c3260e8857b6542802073bb730f6756071b46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c1d53876b7b863c612d597fd716c3260e8857b6542802073bb730f6756071b46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9mrr7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:37Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:37 crc kubenswrapper[4967]: I1008 00:07:37.133820 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:37 crc kubenswrapper[4967]: I1008 00:07:37.133853 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:37 crc kubenswrapper[4967]: I1008 00:07:37.133883 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:37 crc kubenswrapper[4967]: I1008 00:07:37.133899 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:37 crc kubenswrapper[4967]: I1008 00:07:37.133911 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:37Z","lastTransitionTime":"2025-10-08T00:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:37 crc kubenswrapper[4967]: I1008 00:07:37.136388 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kpmlg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f975d47d-42a6-422e-80e2-d1ee282817fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a317ae6640eb38b3aee0f9738cca890e8ab3d92075b620cd312f40a0628d46c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qznqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ef61854aea40924c5372b0553cd0b7e8faa39a1d42fec5b65cf867c368f72c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qznqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-kpmlg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:37Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:37 crc kubenswrapper[4967]: I1008 00:07:37.149240 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58c4704-a88a-43b4-b944-4de626f13f7e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22f5dabb39da0a7bd01db4dae1437b9a231a1fcc0dfe749a4a5a637a72cf33c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01a2b3dc785520c79eae998482e5a1347725c76584357073238fe60a619edb77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6aa6f5e81feceab3d1c7587152a70546b365e629a69267e4d3084324a1de764a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63c2b7c4b2791bd1f16888c58b878dbff1f87884e68c9949de914d0d3e1d39fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:37Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:37 crc kubenswrapper[4967]: I1008 00:07:37.236234 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:37 crc kubenswrapper[4967]: I1008 00:07:37.236269 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:37 crc kubenswrapper[4967]: I1008 00:07:37.236281 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:37 crc kubenswrapper[4967]: I1008 00:07:37.236296 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:37 crc kubenswrapper[4967]: I1008 00:07:37.236308 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:37Z","lastTransitionTime":"2025-10-08T00:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:37 crc kubenswrapper[4967]: I1008 00:07:37.340008 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:37 crc kubenswrapper[4967]: I1008 00:07:37.340066 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:37 crc kubenswrapper[4967]: I1008 00:07:37.340086 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:37 crc kubenswrapper[4967]: I1008 00:07:37.340111 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:37 crc kubenswrapper[4967]: I1008 00:07:37.340128 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:37Z","lastTransitionTime":"2025-10-08T00:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:37 crc kubenswrapper[4967]: I1008 00:07:37.443239 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:37 crc kubenswrapper[4967]: I1008 00:07:37.443313 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:37 crc kubenswrapper[4967]: I1008 00:07:37.443337 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:37 crc kubenswrapper[4967]: I1008 00:07:37.443364 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:37 crc kubenswrapper[4967]: I1008 00:07:37.443385 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:37Z","lastTransitionTime":"2025-10-08T00:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:37 crc kubenswrapper[4967]: I1008 00:07:37.546210 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:37 crc kubenswrapper[4967]: I1008 00:07:37.546280 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:37 crc kubenswrapper[4967]: I1008 00:07:37.546302 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:37 crc kubenswrapper[4967]: I1008 00:07:37.546329 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:37 crc kubenswrapper[4967]: I1008 00:07:37.546389 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:37Z","lastTransitionTime":"2025-10-08T00:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:37 crc kubenswrapper[4967]: I1008 00:07:37.649309 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:37 crc kubenswrapper[4967]: I1008 00:07:37.649399 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:37 crc kubenswrapper[4967]: I1008 00:07:37.649427 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:37 crc kubenswrapper[4967]: I1008 00:07:37.649459 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:37 crc kubenswrapper[4967]: I1008 00:07:37.649484 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:37Z","lastTransitionTime":"2025-10-08T00:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:37 crc kubenswrapper[4967]: I1008 00:07:37.752118 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:37 crc kubenswrapper[4967]: I1008 00:07:37.752179 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:37 crc kubenswrapper[4967]: I1008 00:07:37.752196 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:37 crc kubenswrapper[4967]: I1008 00:07:37.752218 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:37 crc kubenswrapper[4967]: I1008 00:07:37.752238 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:37Z","lastTransitionTime":"2025-10-08T00:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:37 crc kubenswrapper[4967]: I1008 00:07:37.856015 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:37 crc kubenswrapper[4967]: I1008 00:07:37.856051 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:37 crc kubenswrapper[4967]: I1008 00:07:37.856062 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:37 crc kubenswrapper[4967]: I1008 00:07:37.856079 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:37 crc kubenswrapper[4967]: I1008 00:07:37.856091 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:37Z","lastTransitionTime":"2025-10-08T00:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:37 crc kubenswrapper[4967]: I1008 00:07:37.959880 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:37 crc kubenswrapper[4967]: I1008 00:07:37.959941 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:37 crc kubenswrapper[4967]: I1008 00:07:37.959964 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:37 crc kubenswrapper[4967]: I1008 00:07:37.959992 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:37 crc kubenswrapper[4967]: I1008 00:07:37.960012 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:37Z","lastTransitionTime":"2025-10-08T00:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:38 crc kubenswrapper[4967]: I1008 00:07:38.063151 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:38 crc kubenswrapper[4967]: I1008 00:07:38.063231 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:38 crc kubenswrapper[4967]: I1008 00:07:38.063252 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:38 crc kubenswrapper[4967]: I1008 00:07:38.063273 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:38 crc kubenswrapper[4967]: I1008 00:07:38.063285 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:38Z","lastTransitionTime":"2025-10-08T00:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:38 crc kubenswrapper[4967]: I1008 00:07:38.117062 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 00:07:38 crc kubenswrapper[4967]: I1008 00:07:38.117192 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2h7z8" Oct 08 00:07:38 crc kubenswrapper[4967]: E1008 00:07:38.117294 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 00:07:38 crc kubenswrapper[4967]: E1008 00:07:38.117462 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2h7z8" podUID="59f68771-e25d-4367-87ee-68a31d767882" Oct 08 00:07:38 crc kubenswrapper[4967]: I1008 00:07:38.166867 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:38 crc kubenswrapper[4967]: I1008 00:07:38.166935 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:38 crc kubenswrapper[4967]: I1008 00:07:38.166961 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:38 crc kubenswrapper[4967]: I1008 00:07:38.166995 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:38 crc kubenswrapper[4967]: I1008 00:07:38.167019 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:38Z","lastTransitionTime":"2025-10-08T00:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:38 crc kubenswrapper[4967]: I1008 00:07:38.270781 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:38 crc kubenswrapper[4967]: I1008 00:07:38.270879 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:38 crc kubenswrapper[4967]: I1008 00:07:38.270905 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:38 crc kubenswrapper[4967]: I1008 00:07:38.270934 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:38 crc kubenswrapper[4967]: I1008 00:07:38.270957 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:38Z","lastTransitionTime":"2025-10-08T00:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:38 crc kubenswrapper[4967]: I1008 00:07:38.373344 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:38 crc kubenswrapper[4967]: I1008 00:07:38.373402 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:38 crc kubenswrapper[4967]: I1008 00:07:38.373420 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:38 crc kubenswrapper[4967]: I1008 00:07:38.373444 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:38 crc kubenswrapper[4967]: I1008 00:07:38.373461 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:38Z","lastTransitionTime":"2025-10-08T00:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:38 crc kubenswrapper[4967]: I1008 00:07:38.475490 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:38 crc kubenswrapper[4967]: I1008 00:07:38.475527 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:38 crc kubenswrapper[4967]: I1008 00:07:38.475536 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:38 crc kubenswrapper[4967]: I1008 00:07:38.475552 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:38 crc kubenswrapper[4967]: I1008 00:07:38.475561 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:38Z","lastTransitionTime":"2025-10-08T00:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:38 crc kubenswrapper[4967]: I1008 00:07:38.577793 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:38 crc kubenswrapper[4967]: I1008 00:07:38.577851 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:38 crc kubenswrapper[4967]: I1008 00:07:38.577873 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:38 crc kubenswrapper[4967]: I1008 00:07:38.577904 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:38 crc kubenswrapper[4967]: I1008 00:07:38.577926 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:38Z","lastTransitionTime":"2025-10-08T00:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:38 crc kubenswrapper[4967]: I1008 00:07:38.680471 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:38 crc kubenswrapper[4967]: I1008 00:07:38.680506 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:38 crc kubenswrapper[4967]: I1008 00:07:38.680518 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:38 crc kubenswrapper[4967]: I1008 00:07:38.680533 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:38 crc kubenswrapper[4967]: I1008 00:07:38.680544 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:38Z","lastTransitionTime":"2025-10-08T00:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:38 crc kubenswrapper[4967]: I1008 00:07:38.782941 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:38 crc kubenswrapper[4967]: I1008 00:07:38.783020 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:38 crc kubenswrapper[4967]: I1008 00:07:38.783055 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:38 crc kubenswrapper[4967]: I1008 00:07:38.783086 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:38 crc kubenswrapper[4967]: I1008 00:07:38.783109 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:38Z","lastTransitionTime":"2025-10-08T00:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:38 crc kubenswrapper[4967]: I1008 00:07:38.897612 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:38 crc kubenswrapper[4967]: I1008 00:07:38.897686 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:38 crc kubenswrapper[4967]: I1008 00:07:38.897701 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:38 crc kubenswrapper[4967]: I1008 00:07:38.897720 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:38 crc kubenswrapper[4967]: I1008 00:07:38.897731 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:38Z","lastTransitionTime":"2025-10-08T00:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:39 crc kubenswrapper[4967]: I1008 00:07:38.999910 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:39 crc kubenswrapper[4967]: I1008 00:07:38.999946 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:39 crc kubenswrapper[4967]: I1008 00:07:38.999959 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:39 crc kubenswrapper[4967]: I1008 00:07:38.999973 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:39 crc kubenswrapper[4967]: I1008 00:07:38.999981 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:38Z","lastTransitionTime":"2025-10-08T00:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:39 crc kubenswrapper[4967]: I1008 00:07:39.102594 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:39 crc kubenswrapper[4967]: I1008 00:07:39.103020 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:39 crc kubenswrapper[4967]: I1008 00:07:39.103041 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:39 crc kubenswrapper[4967]: I1008 00:07:39.103062 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:39 crc kubenswrapper[4967]: I1008 00:07:39.103074 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:39Z","lastTransitionTime":"2025-10-08T00:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:39 crc kubenswrapper[4967]: I1008 00:07:39.116909 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 00:07:39 crc kubenswrapper[4967]: I1008 00:07:39.116931 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 00:07:39 crc kubenswrapper[4967]: E1008 00:07:39.117054 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 00:07:39 crc kubenswrapper[4967]: E1008 00:07:39.117141 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 00:07:39 crc kubenswrapper[4967]: I1008 00:07:39.205526 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:39 crc kubenswrapper[4967]: I1008 00:07:39.205595 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:39 crc kubenswrapper[4967]: I1008 00:07:39.205609 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:39 crc kubenswrapper[4967]: I1008 00:07:39.205626 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:39 crc kubenswrapper[4967]: I1008 00:07:39.205639 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:39Z","lastTransitionTime":"2025-10-08T00:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:39 crc kubenswrapper[4967]: I1008 00:07:39.308555 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:39 crc kubenswrapper[4967]: I1008 00:07:39.309390 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:39 crc kubenswrapper[4967]: I1008 00:07:39.309468 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:39 crc kubenswrapper[4967]: I1008 00:07:39.309499 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:39 crc kubenswrapper[4967]: I1008 00:07:39.309551 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:39Z","lastTransitionTime":"2025-10-08T00:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:39 crc kubenswrapper[4967]: I1008 00:07:39.412845 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:39 crc kubenswrapper[4967]: I1008 00:07:39.412923 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:39 crc kubenswrapper[4967]: I1008 00:07:39.412941 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:39 crc kubenswrapper[4967]: I1008 00:07:39.412966 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:39 crc kubenswrapper[4967]: I1008 00:07:39.412983 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:39Z","lastTransitionTime":"2025-10-08T00:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:39 crc kubenswrapper[4967]: I1008 00:07:39.434623 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:39 crc kubenswrapper[4967]: I1008 00:07:39.434721 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:39 crc kubenswrapper[4967]: I1008 00:07:39.434741 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:39 crc kubenswrapper[4967]: I1008 00:07:39.434763 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:39 crc kubenswrapper[4967]: I1008 00:07:39.434781 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:39Z","lastTransitionTime":"2025-10-08T00:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:39 crc kubenswrapper[4967]: E1008 00:07:39.451523 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5cddad4c-22d1-42d7-a373-87b47c1cc97b\\\",\\\"systemUUID\\\":\\\"b243ca1c-20cd-4e6f-96b2-52fce74dd2d4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:39Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:39 crc kubenswrapper[4967]: I1008 00:07:39.455136 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:39 crc kubenswrapper[4967]: I1008 00:07:39.455204 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:39 crc kubenswrapper[4967]: I1008 00:07:39.455218 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:39 crc kubenswrapper[4967]: I1008 00:07:39.455236 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:39 crc kubenswrapper[4967]: I1008 00:07:39.455271 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:39Z","lastTransitionTime":"2025-10-08T00:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:39 crc kubenswrapper[4967]: E1008 00:07:39.473398 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5cddad4c-22d1-42d7-a373-87b47c1cc97b\\\",\\\"systemUUID\\\":\\\"b243ca1c-20cd-4e6f-96b2-52fce74dd2d4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:39Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:39 crc kubenswrapper[4967]: I1008 00:07:39.478449 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:39 crc kubenswrapper[4967]: I1008 00:07:39.478508 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:39 crc kubenswrapper[4967]: I1008 00:07:39.478528 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:39 crc kubenswrapper[4967]: I1008 00:07:39.478547 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:39 crc kubenswrapper[4967]: I1008 00:07:39.478560 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:39Z","lastTransitionTime":"2025-10-08T00:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:39 crc kubenswrapper[4967]: E1008 00:07:39.492047 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5cddad4c-22d1-42d7-a373-87b47c1cc97b\\\",\\\"systemUUID\\\":\\\"b243ca1c-20cd-4e6f-96b2-52fce74dd2d4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:39Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:39 crc kubenswrapper[4967]: I1008 00:07:39.495428 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:39 crc kubenswrapper[4967]: I1008 00:07:39.495475 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:39 crc kubenswrapper[4967]: I1008 00:07:39.495492 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:39 crc kubenswrapper[4967]: I1008 00:07:39.495509 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:39 crc kubenswrapper[4967]: I1008 00:07:39.495521 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:39Z","lastTransitionTime":"2025-10-08T00:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:39 crc kubenswrapper[4967]: E1008 00:07:39.506801 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5cddad4c-22d1-42d7-a373-87b47c1cc97b\\\",\\\"systemUUID\\\":\\\"b243ca1c-20cd-4e6f-96b2-52fce74dd2d4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:39Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:39 crc kubenswrapper[4967]: I1008 00:07:39.510994 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:39 crc kubenswrapper[4967]: I1008 00:07:39.511030 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:39 crc kubenswrapper[4967]: I1008 00:07:39.511045 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:39 crc kubenswrapper[4967]: I1008 00:07:39.511062 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:39 crc kubenswrapper[4967]: I1008 00:07:39.511075 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:39Z","lastTransitionTime":"2025-10-08T00:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:39 crc kubenswrapper[4967]: E1008 00:07:39.522695 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5cddad4c-22d1-42d7-a373-87b47c1cc97b\\\",\\\"systemUUID\\\":\\\"b243ca1c-20cd-4e6f-96b2-52fce74dd2d4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:39Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:39 crc kubenswrapper[4967]: E1008 00:07:39.522816 4967 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 08 00:07:39 crc kubenswrapper[4967]: I1008 00:07:39.524515 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:39 crc kubenswrapper[4967]: I1008 00:07:39.524551 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:39 crc kubenswrapper[4967]: I1008 00:07:39.524561 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:39 crc kubenswrapper[4967]: I1008 00:07:39.524609 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:39 crc kubenswrapper[4967]: I1008 00:07:39.524620 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:39Z","lastTransitionTime":"2025-10-08T00:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:39 crc kubenswrapper[4967]: I1008 00:07:39.627342 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:39 crc kubenswrapper[4967]: I1008 00:07:39.627382 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:39 crc kubenswrapper[4967]: I1008 00:07:39.627394 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:39 crc kubenswrapper[4967]: I1008 00:07:39.627409 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:39 crc kubenswrapper[4967]: I1008 00:07:39.627421 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:39Z","lastTransitionTime":"2025-10-08T00:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:39 crc kubenswrapper[4967]: I1008 00:07:39.730039 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:39 crc kubenswrapper[4967]: I1008 00:07:39.730098 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:39 crc kubenswrapper[4967]: I1008 00:07:39.730110 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:39 crc kubenswrapper[4967]: I1008 00:07:39.730130 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:39 crc kubenswrapper[4967]: I1008 00:07:39.730145 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:39Z","lastTransitionTime":"2025-10-08T00:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:39 crc kubenswrapper[4967]: I1008 00:07:39.833533 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:39 crc kubenswrapper[4967]: I1008 00:07:39.833611 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:39 crc kubenswrapper[4967]: I1008 00:07:39.833648 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:39 crc kubenswrapper[4967]: I1008 00:07:39.833723 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:39 crc kubenswrapper[4967]: I1008 00:07:39.833748 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:39Z","lastTransitionTime":"2025-10-08T00:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:39 crc kubenswrapper[4967]: I1008 00:07:39.936392 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:39 crc kubenswrapper[4967]: I1008 00:07:39.936462 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:39 crc kubenswrapper[4967]: I1008 00:07:39.936477 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:39 crc kubenswrapper[4967]: I1008 00:07:39.936506 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:39 crc kubenswrapper[4967]: I1008 00:07:39.936527 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:39Z","lastTransitionTime":"2025-10-08T00:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:40 crc kubenswrapper[4967]: I1008 00:07:40.039046 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:40 crc kubenswrapper[4967]: I1008 00:07:40.039197 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:40 crc kubenswrapper[4967]: I1008 00:07:40.039217 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:40 crc kubenswrapper[4967]: I1008 00:07:40.039240 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:40 crc kubenswrapper[4967]: I1008 00:07:40.039257 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:40Z","lastTransitionTime":"2025-10-08T00:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:40 crc kubenswrapper[4967]: I1008 00:07:40.117399 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2h7z8" Oct 08 00:07:40 crc kubenswrapper[4967]: I1008 00:07:40.117487 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 00:07:40 crc kubenswrapper[4967]: E1008 00:07:40.117586 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2h7z8" podUID="59f68771-e25d-4367-87ee-68a31d767882" Oct 08 00:07:40 crc kubenswrapper[4967]: E1008 00:07:40.117637 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 00:07:40 crc kubenswrapper[4967]: I1008 00:07:40.144222 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:40 crc kubenswrapper[4967]: I1008 00:07:40.144290 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:40 crc kubenswrapper[4967]: I1008 00:07:40.144306 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:40 crc kubenswrapper[4967]: I1008 00:07:40.144329 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:40 crc kubenswrapper[4967]: I1008 00:07:40.144346 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:40Z","lastTransitionTime":"2025-10-08T00:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:40 crc kubenswrapper[4967]: I1008 00:07:40.246976 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:40 crc kubenswrapper[4967]: I1008 00:07:40.247023 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:40 crc kubenswrapper[4967]: I1008 00:07:40.247035 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:40 crc kubenswrapper[4967]: I1008 00:07:40.247052 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:40 crc kubenswrapper[4967]: I1008 00:07:40.247065 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:40Z","lastTransitionTime":"2025-10-08T00:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:40 crc kubenswrapper[4967]: I1008 00:07:40.349883 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:40 crc kubenswrapper[4967]: I1008 00:07:40.349935 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:40 crc kubenswrapper[4967]: I1008 00:07:40.349947 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:40 crc kubenswrapper[4967]: I1008 00:07:40.349965 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:40 crc kubenswrapper[4967]: I1008 00:07:40.349976 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:40Z","lastTransitionTime":"2025-10-08T00:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:40 crc kubenswrapper[4967]: I1008 00:07:40.452769 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:40 crc kubenswrapper[4967]: I1008 00:07:40.452816 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:40 crc kubenswrapper[4967]: I1008 00:07:40.452827 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:40 crc kubenswrapper[4967]: I1008 00:07:40.452845 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:40 crc kubenswrapper[4967]: I1008 00:07:40.452864 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:40Z","lastTransitionTime":"2025-10-08T00:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:40 crc kubenswrapper[4967]: I1008 00:07:40.555497 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:40 crc kubenswrapper[4967]: I1008 00:07:40.555533 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:40 crc kubenswrapper[4967]: I1008 00:07:40.555544 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:40 crc kubenswrapper[4967]: I1008 00:07:40.555558 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:40 crc kubenswrapper[4967]: I1008 00:07:40.555567 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:40Z","lastTransitionTime":"2025-10-08T00:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:40 crc kubenswrapper[4967]: I1008 00:07:40.657703 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:40 crc kubenswrapper[4967]: I1008 00:07:40.657738 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:40 crc kubenswrapper[4967]: I1008 00:07:40.657749 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:40 crc kubenswrapper[4967]: I1008 00:07:40.657762 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:40 crc kubenswrapper[4967]: I1008 00:07:40.657771 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:40Z","lastTransitionTime":"2025-10-08T00:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:40 crc kubenswrapper[4967]: I1008 00:07:40.760950 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:40 crc kubenswrapper[4967]: I1008 00:07:40.761035 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:40 crc kubenswrapper[4967]: I1008 00:07:40.761062 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:40 crc kubenswrapper[4967]: I1008 00:07:40.761095 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:40 crc kubenswrapper[4967]: I1008 00:07:40.761118 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:40Z","lastTransitionTime":"2025-10-08T00:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:40 crc kubenswrapper[4967]: I1008 00:07:40.864061 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:40 crc kubenswrapper[4967]: I1008 00:07:40.864119 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:40 crc kubenswrapper[4967]: I1008 00:07:40.864135 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:40 crc kubenswrapper[4967]: I1008 00:07:40.864155 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:40 crc kubenswrapper[4967]: I1008 00:07:40.864169 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:40Z","lastTransitionTime":"2025-10-08T00:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:40 crc kubenswrapper[4967]: I1008 00:07:40.966623 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:40 crc kubenswrapper[4967]: I1008 00:07:40.966661 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:40 crc kubenswrapper[4967]: I1008 00:07:40.966700 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:40 crc kubenswrapper[4967]: I1008 00:07:40.966721 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:40 crc kubenswrapper[4967]: I1008 00:07:40.966735 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:40Z","lastTransitionTime":"2025-10-08T00:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:41 crc kubenswrapper[4967]: I1008 00:07:41.069445 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:41 crc kubenswrapper[4967]: I1008 00:07:41.069521 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:41 crc kubenswrapper[4967]: I1008 00:07:41.069541 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:41 crc kubenswrapper[4967]: I1008 00:07:41.069567 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:41 crc kubenswrapper[4967]: I1008 00:07:41.069587 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:41Z","lastTransitionTime":"2025-10-08T00:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:41 crc kubenswrapper[4967]: I1008 00:07:41.117715 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 00:07:41 crc kubenswrapper[4967]: E1008 00:07:41.117911 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 00:07:41 crc kubenswrapper[4967]: I1008 00:07:41.117933 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 00:07:41 crc kubenswrapper[4967]: E1008 00:07:41.118446 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 00:07:41 crc kubenswrapper[4967]: I1008 00:07:41.118963 4967 scope.go:117] "RemoveContainer" containerID="d1cfd26b02d0e17b111b8ad84dab6c2adf692de66ff77158ee7cbff941c66ecd" Oct 08 00:07:41 crc kubenswrapper[4967]: E1008 00:07:41.119253 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-bs8w9_openshift-ovn-kubernetes(4f1af2ba-8907-4d14-afa1-ffde207c80f4)\"" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" podUID="4f1af2ba-8907-4d14-afa1-ffde207c80f4" Oct 08 00:07:41 crc kubenswrapper[4967]: I1008 00:07:41.172705 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:41 crc kubenswrapper[4967]: I1008 00:07:41.172761 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:41 crc kubenswrapper[4967]: I1008 00:07:41.172779 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:41 crc kubenswrapper[4967]: I1008 00:07:41.172804 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:41 crc kubenswrapper[4967]: I1008 00:07:41.172818 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:41Z","lastTransitionTime":"2025-10-08T00:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:41 crc kubenswrapper[4967]: I1008 00:07:41.277138 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:41 crc kubenswrapper[4967]: I1008 00:07:41.277221 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:41 crc kubenswrapper[4967]: I1008 00:07:41.277242 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:41 crc kubenswrapper[4967]: I1008 00:07:41.277280 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:41 crc kubenswrapper[4967]: I1008 00:07:41.277326 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:41Z","lastTransitionTime":"2025-10-08T00:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:41 crc kubenswrapper[4967]: I1008 00:07:41.381322 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:41 crc kubenswrapper[4967]: I1008 00:07:41.381403 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:41 crc kubenswrapper[4967]: I1008 00:07:41.381415 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:41 crc kubenswrapper[4967]: I1008 00:07:41.381465 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:41 crc kubenswrapper[4967]: I1008 00:07:41.381483 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:41Z","lastTransitionTime":"2025-10-08T00:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:41 crc kubenswrapper[4967]: I1008 00:07:41.484308 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:41 crc kubenswrapper[4967]: I1008 00:07:41.484373 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:41 crc kubenswrapper[4967]: I1008 00:07:41.484388 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:41 crc kubenswrapper[4967]: I1008 00:07:41.484415 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:41 crc kubenswrapper[4967]: I1008 00:07:41.484433 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:41Z","lastTransitionTime":"2025-10-08T00:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:41 crc kubenswrapper[4967]: I1008 00:07:41.586806 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:41 crc kubenswrapper[4967]: I1008 00:07:41.586857 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:41 crc kubenswrapper[4967]: I1008 00:07:41.586868 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:41 crc kubenswrapper[4967]: I1008 00:07:41.586886 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:41 crc kubenswrapper[4967]: I1008 00:07:41.586899 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:41Z","lastTransitionTime":"2025-10-08T00:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:41 crc kubenswrapper[4967]: I1008 00:07:41.690335 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:41 crc kubenswrapper[4967]: I1008 00:07:41.690412 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:41 crc kubenswrapper[4967]: I1008 00:07:41.690426 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:41 crc kubenswrapper[4967]: I1008 00:07:41.690446 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:41 crc kubenswrapper[4967]: I1008 00:07:41.690458 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:41Z","lastTransitionTime":"2025-10-08T00:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:41 crc kubenswrapper[4967]: I1008 00:07:41.794031 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:41 crc kubenswrapper[4967]: I1008 00:07:41.794083 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:41 crc kubenswrapper[4967]: I1008 00:07:41.794099 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:41 crc kubenswrapper[4967]: I1008 00:07:41.794118 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:41 crc kubenswrapper[4967]: I1008 00:07:41.794131 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:41Z","lastTransitionTime":"2025-10-08T00:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:41 crc kubenswrapper[4967]: I1008 00:07:41.897269 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:41 crc kubenswrapper[4967]: I1008 00:07:41.897308 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:41 crc kubenswrapper[4967]: I1008 00:07:41.897319 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:41 crc kubenswrapper[4967]: I1008 00:07:41.897334 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:41 crc kubenswrapper[4967]: I1008 00:07:41.897346 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:41Z","lastTransitionTime":"2025-10-08T00:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:41 crc kubenswrapper[4967]: I1008 00:07:41.999435 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:41 crc kubenswrapper[4967]: I1008 00:07:41.999477 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:41 crc kubenswrapper[4967]: I1008 00:07:41.999493 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:41 crc kubenswrapper[4967]: I1008 00:07:41.999509 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:41 crc kubenswrapper[4967]: I1008 00:07:41.999522 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:41Z","lastTransitionTime":"2025-10-08T00:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:42 crc kubenswrapper[4967]: I1008 00:07:42.101819 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:42 crc kubenswrapper[4967]: I1008 00:07:42.101957 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:42 crc kubenswrapper[4967]: I1008 00:07:42.101979 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:42 crc kubenswrapper[4967]: I1008 00:07:42.102002 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:42 crc kubenswrapper[4967]: I1008 00:07:42.102019 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:42Z","lastTransitionTime":"2025-10-08T00:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:42 crc kubenswrapper[4967]: I1008 00:07:42.117573 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 00:07:42 crc kubenswrapper[4967]: I1008 00:07:42.117607 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2h7z8" Oct 08 00:07:42 crc kubenswrapper[4967]: E1008 00:07:42.117941 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2h7z8" podUID="59f68771-e25d-4367-87ee-68a31d767882" Oct 08 00:07:42 crc kubenswrapper[4967]: E1008 00:07:42.118603 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 00:07:42 crc kubenswrapper[4967]: I1008 00:07:42.132040 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94e63bc5641ce80d507f76a19b203f78c42368a7a4105f3fd09e19af21a1b735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:42Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:42 crc kubenswrapper[4967]: I1008 00:07:42.146990 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e98666fd3152bab57d3723e7da72aae2703b79f7ca73eb3c3e5f0e062a82592\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cf1cbe5548146fe359aafa351864125ec6b8bdd7b796aa4d311f6f3a16ad212\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:42Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:42 crc kubenswrapper[4967]: I1008 00:07:42.158098 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9tdbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41264a12-9486-4fee-a08c-607b590e5a70\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://806e85df796ae300d279f73a0cc4433a5bb1f7620ee8ce33c8dc4674c4ce0596\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5blv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9tdbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:42Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:42 crc kubenswrapper[4967]: I1008 00:07:42.177047 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f1af2ba-8907-4d14-afa1-ffde207c80f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4535cffb8c9d64e662e282ecb44f6817b4604866f0bb75015f3b888d5554f821\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b03b83ddff429dbfa8fe052901aaa86a39ff0dd75477ffa8a6b6f4198f5819ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c9fa0559eb35ba67e11e36069ac8d0b2fec76399788eef668a08ad4353df3ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3312263e12bd42fbed0476ab5ace66b5497f62c9477865e130d67e3922e05cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e69dc603055aa221d44d11c049ddcc50cec568979911a310ed773a90341a499\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115f7a33f3dedae8f59d2f45f465f95f48b99c824e895ef7a0be9fe7a1b61d23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1cfd26b02d0e17b111b8ad84dab6c2adf692de66ff77158ee7cbff941c66ecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1cfd26b02d0e17b111b8ad84dab6c2adf692de66ff77158ee7cbff941c66ecd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T00:07:29Z\\\",\\\"message\\\":\\\"iled objects of type *v1.Pod\\\\nI1008 00:07:28.973629 6638 obj_retry.go:409] Going to retry *v1.Pod resource setup for 1 objects: [openshift-multus/network-metrics-daemon-2h7z8]\\\\nI1008 00:07:28.973646 6638 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1008 00:07:28.973685 6638 obj_retry.go:285] Attempting retry of *v1.Pod openshift-multus/network-metrics-daemon-2h7z8 before timer (time: 2025-10-08 00:07:30.377116097 +0000 UTC m=+1.948919645): skip\\\\nI1008 00:07:28.973706 6638 obj_retry.go:420] Function iterateRetryResources for *v1.Pod ended (in 79.172µs)\\\\nI1008 00:07:28.973904 6638 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1008 00:07:28.973958 6638 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1008 00:07:28.974084 6638 factory.go:656] Stopping watch factory\\\\nI1008 00:07:28.974107 6638 ovnkube.go:599] Stopped ovnkube\\\\nI1008 00:07:28.974137 6638 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1008 00:07:28.974149 6638 handler.go:208] Removed *v1.Node event handler 2\\\\nI1008 00:07:28.974163 6638 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1008 00:07:28.974267 6638 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-bs8w9_openshift-ovn-kubernetes(4f1af2ba-8907-4d14-afa1-ffde207c80f4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27fb2b1225e386dcd7fecaaa0aa54ba1fdfafabfde6c3824ae07315313067542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bs8w9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:42Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:42 crc kubenswrapper[4967]: I1008 00:07:42.188314 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-v5n7t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abe7ae2d-7e63-4249-96e1-4000e9628ccc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52575e853ff6727dde7c89e4add4362be126127db27191c2a7bf2871efa719d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpsjd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:06Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-v5n7t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:42Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:42 crc kubenswrapper[4967]: I1008 00:07:42.200116 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6ea2440-b2ae-481e-9442-0b3dda9688ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37fd5dcdfb864320d4d17729ab48c6da505fd867f384ad70a6dcf5bed78a3978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c7893dfaa0370f159d41c30cd869bb18550a819e8c3d20dd6f57dcbcc697a57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c430979d298477b4394c3567bb3800ab7dea874cee8fa41a6a50a19a7e3a0133\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://426e86ab409776ff0fa9505dc0d8ec36f27a5bf21d3aedafb19d3935bdba7529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://426e86ab409776ff0fa9505dc0d8ec36f27a5bf21d3aedafb19d3935bdba7529\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:42Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:42 crc kubenswrapper[4967]: I1008 00:07:42.204950 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:42 crc kubenswrapper[4967]: I1008 00:07:42.205006 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:42 crc kubenswrapper[4967]: I1008 00:07:42.205024 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:42 crc kubenswrapper[4967]: I1008 00:07:42.205048 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:42 crc kubenswrapper[4967]: I1008 00:07:42.205064 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:42Z","lastTransitionTime":"2025-10-08T00:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:42 crc kubenswrapper[4967]: I1008 00:07:42.212542 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:42Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:42 crc kubenswrapper[4967]: I1008 00:07:42.226631 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-npz2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c71501b1-aa9e-420e-b1bd-7e15365e7adc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dfe6c420d804b2d7daef4b5cfe3f87b77b369287c4d65553fa93f5f24f589d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gblk2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-npz2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:42Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:42 crc kubenswrapper[4967]: I1008 00:07:42.236966 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2h7z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59f68771-e25d-4367-87ee-68a31d767882\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7r2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7r2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:18Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2h7z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:42Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:42 crc kubenswrapper[4967]: I1008 00:07:42.254879 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58c4704-a88a-43b4-b944-4de626f13f7e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22f5dabb39da0a7bd01db4dae1437b9a231a1fcc0dfe749a4a5a637a72cf33c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01a2b3dc785520c79eae998482e5a1347725c76584357073238fe60a619edb77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6aa6f5e81feceab3d1c7587152a70546b365e629a69267e4d3084324a1de764a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63c2b7c4b2791bd1f16888c58b878dbff1f87884e68c9949de914d0d3e1d39fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:42Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:42 crc kubenswrapper[4967]: I1008 00:07:42.270406 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:42Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:42 crc kubenswrapper[4967]: I1008 00:07:42.288398 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9mrr7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9299bc4f-f8d3-4baf-9885-39573982fd64\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://994beafece9dead0d85ae74dc453a142460705fd66b52ba22f2fa799cf27da08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb49c8476b482416588c4d403016940730064a619c0641eb029c947a4bab1215\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb49c8476b482416588c4d403016940730064a619c0641eb029c947a4bab1215\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b7b63962dcbf477f1fa4d71826a4aa3358920a2c1b2f5e48716411b5bdca049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b7b63962dcbf477f1fa4d71826a4aa3358920a2c1b2f5e48716411b5bdca049\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1e770dc737173fd4f3220ed6c52b422ccbd829722386a37029e1c3e10d80e1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1e770dc737173fd4f3220ed6c52b422ccbd829722386a37029e1c3e10d80e1a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e357d970878c996f219a79f1876b363176c9847b9575f33090f62bc299529a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e357d970878c996f219a79f1876b363176c9847b9575f33090f62bc299529a95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f098966bfad49d06278eaa72be042596d488115e597737d9a0a0cfc78f99f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49f098966bfad49d06278eaa72be042596d488115e597737d9a0a0cfc78f99f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1d53876b7b863c612d597fd716c3260e8857b6542802073bb730f6756071b46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c1d53876b7b863c612d597fd716c3260e8857b6542802073bb730f6756071b46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9mrr7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:42Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:42 crc kubenswrapper[4967]: I1008 00:07:42.301785 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kpmlg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f975d47d-42a6-422e-80e2-d1ee282817fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a317ae6640eb38b3aee0f9738cca890e8ab3d92075b620cd312f40a0628d46c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qznqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ef61854aea40924c5372b0553cd0b7e8faa39a1d42fec5b65cf867c368f72c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qznqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-kpmlg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:42Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:42 crc kubenswrapper[4967]: I1008 00:07:42.308272 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:42 crc kubenswrapper[4967]: I1008 00:07:42.308370 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:42 crc kubenswrapper[4967]: I1008 00:07:42.308386 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:42 crc kubenswrapper[4967]: I1008 00:07:42.308413 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:42 crc kubenswrapper[4967]: I1008 00:07:42.308429 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:42Z","lastTransitionTime":"2025-10-08T00:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:42 crc kubenswrapper[4967]: I1008 00:07:42.320862 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fa4b747-ab6e-4a11-9903-ef2dcb9668c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbadcdade274b6d9be940f22bc949e3c55cadf95dcad43c8cb7a1085c0e6708c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d0a0b2ef4a98ee828688081a74f4b7ecc31c8fae415d7257a560ecd43960ce3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96b53b1b283025b3d078a3f505d469d1872ad4bb8ab8e612c36c70b0b00a920\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef86a2422616e1e0d5c3afe844818c2bdf2150f941d1aefc506995f58587b2bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1400e110997c64fa33d6b8672c7e91cab5f782952ddf8e1dea2a882689026e7f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 00:06:56.760088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 00:06:56.761210 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-369805205/tls.crt::/tmp/serving-cert-369805205/tls.key\\\\\\\"\\\\nI1008 00:07:02.321448 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 00:07:02.325932 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 00:07:02.325980 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 00:07:02.326025 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 00:07:02.326042 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 00:07:02.336203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 00:07:02.336240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 00:07:02.336245 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 00:07:02.336250 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 00:07:02.336253 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 00:07:02.336256 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 00:07:02.336258 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 00:07:02.336430 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 00:07:02.340774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:46Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca28d8db19036e47a7ef3dadb70941f6c6cf5735e7703977d036185c46fce40\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5b222ad84c31d8746dfe03124c7287b401deb3ca2c62ce2b4fbd40e638cee0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5b222ad84c31d8746dfe03124c7287b401deb3ca2c62ce2b4fbd40e638cee0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:42Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:42 crc kubenswrapper[4967]: I1008 00:07:42.337698 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:42Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:42 crc kubenswrapper[4967]: I1008 00:07:42.356031 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbfcdc882e2be1cb9cf666c9e7775671078f7dbbfe8421dd15db737b89920a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:42Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:42 crc kubenswrapper[4967]: I1008 00:07:42.370706 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9461162a-c27d-465e-9ca7-2d3936877cd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9778d78ddc7325abd4d77bc5e71b0b114f50be5dda4a438c7cb6b05640c8a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z9tp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28760af04ef321d0f91b2bce3ad074d44b63dabf676cc37f3be6aa8726341a2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z9tp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xvdtp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:42Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:42 crc kubenswrapper[4967]: I1008 00:07:42.411244 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:42 crc kubenswrapper[4967]: I1008 00:07:42.411552 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:42 crc kubenswrapper[4967]: I1008 00:07:42.411657 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:42 crc kubenswrapper[4967]: I1008 00:07:42.411804 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:42 crc kubenswrapper[4967]: I1008 00:07:42.411965 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:42Z","lastTransitionTime":"2025-10-08T00:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:42 crc kubenswrapper[4967]: I1008 00:07:42.515941 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:42 crc kubenswrapper[4967]: I1008 00:07:42.516315 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:42 crc kubenswrapper[4967]: I1008 00:07:42.516425 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:42 crc kubenswrapper[4967]: I1008 00:07:42.516545 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:42 crc kubenswrapper[4967]: I1008 00:07:42.516709 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:42Z","lastTransitionTime":"2025-10-08T00:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:42 crc kubenswrapper[4967]: I1008 00:07:42.620258 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:42 crc kubenswrapper[4967]: I1008 00:07:42.620342 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:42 crc kubenswrapper[4967]: I1008 00:07:42.620364 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:42 crc kubenswrapper[4967]: I1008 00:07:42.620397 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:42 crc kubenswrapper[4967]: I1008 00:07:42.620418 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:42Z","lastTransitionTime":"2025-10-08T00:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:42 crc kubenswrapper[4967]: I1008 00:07:42.723929 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:42 crc kubenswrapper[4967]: I1008 00:07:42.724134 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:42 crc kubenswrapper[4967]: I1008 00:07:42.724162 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:42 crc kubenswrapper[4967]: I1008 00:07:42.724192 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:42 crc kubenswrapper[4967]: I1008 00:07:42.724214 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:42Z","lastTransitionTime":"2025-10-08T00:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:42 crc kubenswrapper[4967]: I1008 00:07:42.827066 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:42 crc kubenswrapper[4967]: I1008 00:07:42.827104 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:42 crc kubenswrapper[4967]: I1008 00:07:42.827114 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:42 crc kubenswrapper[4967]: I1008 00:07:42.827127 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:42 crc kubenswrapper[4967]: I1008 00:07:42.827137 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:42Z","lastTransitionTime":"2025-10-08T00:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:42 crc kubenswrapper[4967]: I1008 00:07:42.929372 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:42 crc kubenswrapper[4967]: I1008 00:07:42.929412 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:42 crc kubenswrapper[4967]: I1008 00:07:42.929425 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:42 crc kubenswrapper[4967]: I1008 00:07:42.929442 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:42 crc kubenswrapper[4967]: I1008 00:07:42.929454 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:42Z","lastTransitionTime":"2025-10-08T00:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:43 crc kubenswrapper[4967]: I1008 00:07:43.039351 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:43 crc kubenswrapper[4967]: I1008 00:07:43.039386 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:43 crc kubenswrapper[4967]: I1008 00:07:43.039398 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:43 crc kubenswrapper[4967]: I1008 00:07:43.039413 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:43 crc kubenswrapper[4967]: I1008 00:07:43.039423 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:43Z","lastTransitionTime":"2025-10-08T00:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:43 crc kubenswrapper[4967]: I1008 00:07:43.117319 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 00:07:43 crc kubenswrapper[4967]: I1008 00:07:43.117390 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 00:07:43 crc kubenswrapper[4967]: E1008 00:07:43.117447 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 00:07:43 crc kubenswrapper[4967]: E1008 00:07:43.117524 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 00:07:43 crc kubenswrapper[4967]: I1008 00:07:43.142905 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:43 crc kubenswrapper[4967]: I1008 00:07:43.143270 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:43 crc kubenswrapper[4967]: I1008 00:07:43.143468 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:43 crc kubenswrapper[4967]: I1008 00:07:43.143646 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:43 crc kubenswrapper[4967]: I1008 00:07:43.143894 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:43Z","lastTransitionTime":"2025-10-08T00:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:43 crc kubenswrapper[4967]: I1008 00:07:43.247907 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:43 crc kubenswrapper[4967]: I1008 00:07:43.247954 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:43 crc kubenswrapper[4967]: I1008 00:07:43.247966 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:43 crc kubenswrapper[4967]: I1008 00:07:43.247983 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:43 crc kubenswrapper[4967]: I1008 00:07:43.247995 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:43Z","lastTransitionTime":"2025-10-08T00:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:43 crc kubenswrapper[4967]: I1008 00:07:43.350499 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:43 crc kubenswrapper[4967]: I1008 00:07:43.350542 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:43 crc kubenswrapper[4967]: I1008 00:07:43.350553 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:43 crc kubenswrapper[4967]: I1008 00:07:43.350569 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:43 crc kubenswrapper[4967]: I1008 00:07:43.350581 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:43Z","lastTransitionTime":"2025-10-08T00:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:43 crc kubenswrapper[4967]: I1008 00:07:43.453359 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:43 crc kubenswrapper[4967]: I1008 00:07:43.453399 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:43 crc kubenswrapper[4967]: I1008 00:07:43.453411 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:43 crc kubenswrapper[4967]: I1008 00:07:43.453428 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:43 crc kubenswrapper[4967]: I1008 00:07:43.453442 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:43Z","lastTransitionTime":"2025-10-08T00:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:43 crc kubenswrapper[4967]: I1008 00:07:43.555762 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:43 crc kubenswrapper[4967]: I1008 00:07:43.555804 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:43 crc kubenswrapper[4967]: I1008 00:07:43.555815 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:43 crc kubenswrapper[4967]: I1008 00:07:43.555831 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:43 crc kubenswrapper[4967]: I1008 00:07:43.555842 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:43Z","lastTransitionTime":"2025-10-08T00:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:43 crc kubenswrapper[4967]: I1008 00:07:43.658279 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:43 crc kubenswrapper[4967]: I1008 00:07:43.658757 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:43 crc kubenswrapper[4967]: I1008 00:07:43.658954 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:43 crc kubenswrapper[4967]: I1008 00:07:43.659135 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:43 crc kubenswrapper[4967]: I1008 00:07:43.659329 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:43Z","lastTransitionTime":"2025-10-08T00:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:43 crc kubenswrapper[4967]: I1008 00:07:43.761754 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:43 crc kubenswrapper[4967]: I1008 00:07:43.762073 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:43 crc kubenswrapper[4967]: I1008 00:07:43.762206 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:43 crc kubenswrapper[4967]: I1008 00:07:43.762329 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:43 crc kubenswrapper[4967]: I1008 00:07:43.762433 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:43Z","lastTransitionTime":"2025-10-08T00:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:43 crc kubenswrapper[4967]: I1008 00:07:43.865081 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:43 crc kubenswrapper[4967]: I1008 00:07:43.865317 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:43 crc kubenswrapper[4967]: I1008 00:07:43.865401 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:43 crc kubenswrapper[4967]: I1008 00:07:43.865512 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:43 crc kubenswrapper[4967]: I1008 00:07:43.865591 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:43Z","lastTransitionTime":"2025-10-08T00:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:43 crc kubenswrapper[4967]: I1008 00:07:43.968418 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:43 crc kubenswrapper[4967]: I1008 00:07:43.968655 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:43 crc kubenswrapper[4967]: I1008 00:07:43.968736 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:43 crc kubenswrapper[4967]: I1008 00:07:43.968771 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:43 crc kubenswrapper[4967]: I1008 00:07:43.968797 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:43Z","lastTransitionTime":"2025-10-08T00:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:44 crc kubenswrapper[4967]: I1008 00:07:44.071951 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:44 crc kubenswrapper[4967]: I1008 00:07:44.071984 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:44 crc kubenswrapper[4967]: I1008 00:07:44.071993 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:44 crc kubenswrapper[4967]: I1008 00:07:44.072007 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:44 crc kubenswrapper[4967]: I1008 00:07:44.072016 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:44Z","lastTransitionTime":"2025-10-08T00:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:44 crc kubenswrapper[4967]: I1008 00:07:44.117619 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2h7z8" Oct 08 00:07:44 crc kubenswrapper[4967]: I1008 00:07:44.117761 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 00:07:44 crc kubenswrapper[4967]: E1008 00:07:44.117957 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2h7z8" podUID="59f68771-e25d-4367-87ee-68a31d767882" Oct 08 00:07:44 crc kubenswrapper[4967]: E1008 00:07:44.118725 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 00:07:44 crc kubenswrapper[4967]: I1008 00:07:44.174270 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:44 crc kubenswrapper[4967]: I1008 00:07:44.174300 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:44 crc kubenswrapper[4967]: I1008 00:07:44.174310 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:44 crc kubenswrapper[4967]: I1008 00:07:44.174324 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:44 crc kubenswrapper[4967]: I1008 00:07:44.174334 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:44Z","lastTransitionTime":"2025-10-08T00:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:44 crc kubenswrapper[4967]: I1008 00:07:44.278282 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:44 crc kubenswrapper[4967]: I1008 00:07:44.278357 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:44 crc kubenswrapper[4967]: I1008 00:07:44.278381 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:44 crc kubenswrapper[4967]: I1008 00:07:44.278410 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:44 crc kubenswrapper[4967]: I1008 00:07:44.278434 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:44Z","lastTransitionTime":"2025-10-08T00:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:44 crc kubenswrapper[4967]: I1008 00:07:44.380819 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:44 crc kubenswrapper[4967]: I1008 00:07:44.380863 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:44 crc kubenswrapper[4967]: I1008 00:07:44.380987 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:44 crc kubenswrapper[4967]: I1008 00:07:44.381006 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:44 crc kubenswrapper[4967]: I1008 00:07:44.381018 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:44Z","lastTransitionTime":"2025-10-08T00:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:44 crc kubenswrapper[4967]: I1008 00:07:44.483394 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:44 crc kubenswrapper[4967]: I1008 00:07:44.483491 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:44 crc kubenswrapper[4967]: I1008 00:07:44.483521 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:44 crc kubenswrapper[4967]: I1008 00:07:44.483552 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:44 crc kubenswrapper[4967]: I1008 00:07:44.483569 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:44Z","lastTransitionTime":"2025-10-08T00:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:44 crc kubenswrapper[4967]: I1008 00:07:44.586033 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:44 crc kubenswrapper[4967]: I1008 00:07:44.586103 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:44 crc kubenswrapper[4967]: I1008 00:07:44.586129 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:44 crc kubenswrapper[4967]: I1008 00:07:44.586161 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:44 crc kubenswrapper[4967]: I1008 00:07:44.586181 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:44Z","lastTransitionTime":"2025-10-08T00:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:44 crc kubenswrapper[4967]: I1008 00:07:44.689540 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:44 crc kubenswrapper[4967]: I1008 00:07:44.689645 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:44 crc kubenswrapper[4967]: I1008 00:07:44.689690 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:44 crc kubenswrapper[4967]: I1008 00:07:44.689727 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:44 crc kubenswrapper[4967]: I1008 00:07:44.689746 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:44Z","lastTransitionTime":"2025-10-08T00:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:44 crc kubenswrapper[4967]: I1008 00:07:44.792934 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:44 crc kubenswrapper[4967]: I1008 00:07:44.793023 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:44 crc kubenswrapper[4967]: I1008 00:07:44.793059 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:44 crc kubenswrapper[4967]: I1008 00:07:44.793094 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:44 crc kubenswrapper[4967]: I1008 00:07:44.793126 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:44Z","lastTransitionTime":"2025-10-08T00:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:44 crc kubenswrapper[4967]: I1008 00:07:44.895658 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:44 crc kubenswrapper[4967]: I1008 00:07:44.895730 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:44 crc kubenswrapper[4967]: I1008 00:07:44.895742 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:44 crc kubenswrapper[4967]: I1008 00:07:44.895760 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:44 crc kubenswrapper[4967]: I1008 00:07:44.895776 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:44Z","lastTransitionTime":"2025-10-08T00:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:44 crc kubenswrapper[4967]: I1008 00:07:44.999228 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:44 crc kubenswrapper[4967]: I1008 00:07:44.999274 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:44 crc kubenswrapper[4967]: I1008 00:07:44.999286 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:44 crc kubenswrapper[4967]: I1008 00:07:44.999303 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:44 crc kubenswrapper[4967]: I1008 00:07:44.999318 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:44Z","lastTransitionTime":"2025-10-08T00:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:45 crc kubenswrapper[4967]: I1008 00:07:45.102802 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:45 crc kubenswrapper[4967]: I1008 00:07:45.103480 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:45 crc kubenswrapper[4967]: I1008 00:07:45.103507 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:45 crc kubenswrapper[4967]: I1008 00:07:45.103537 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:45 crc kubenswrapper[4967]: I1008 00:07:45.103558 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:45Z","lastTransitionTime":"2025-10-08T00:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:45 crc kubenswrapper[4967]: I1008 00:07:45.117446 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 00:07:45 crc kubenswrapper[4967]: I1008 00:07:45.117512 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 00:07:45 crc kubenswrapper[4967]: E1008 00:07:45.117620 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 00:07:45 crc kubenswrapper[4967]: E1008 00:07:45.117864 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 00:07:45 crc kubenswrapper[4967]: I1008 00:07:45.206034 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:45 crc kubenswrapper[4967]: I1008 00:07:45.206102 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:45 crc kubenswrapper[4967]: I1008 00:07:45.206118 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:45 crc kubenswrapper[4967]: I1008 00:07:45.206141 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:45 crc kubenswrapper[4967]: I1008 00:07:45.206157 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:45Z","lastTransitionTime":"2025-10-08T00:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:45 crc kubenswrapper[4967]: I1008 00:07:45.320557 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:45 crc kubenswrapper[4967]: I1008 00:07:45.320598 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:45 crc kubenswrapper[4967]: I1008 00:07:45.320609 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:45 crc kubenswrapper[4967]: I1008 00:07:45.320629 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:45 crc kubenswrapper[4967]: I1008 00:07:45.320639 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:45Z","lastTransitionTime":"2025-10-08T00:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:45 crc kubenswrapper[4967]: I1008 00:07:45.423316 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:45 crc kubenswrapper[4967]: I1008 00:07:45.423351 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:45 crc kubenswrapper[4967]: I1008 00:07:45.423360 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:45 crc kubenswrapper[4967]: I1008 00:07:45.423372 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:45 crc kubenswrapper[4967]: I1008 00:07:45.423381 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:45Z","lastTransitionTime":"2025-10-08T00:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:45 crc kubenswrapper[4967]: I1008 00:07:45.525886 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:45 crc kubenswrapper[4967]: I1008 00:07:45.525930 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:45 crc kubenswrapper[4967]: I1008 00:07:45.525942 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:45 crc kubenswrapper[4967]: I1008 00:07:45.525957 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:45 crc kubenswrapper[4967]: I1008 00:07:45.525966 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:45Z","lastTransitionTime":"2025-10-08T00:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:45 crc kubenswrapper[4967]: I1008 00:07:45.628359 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:45 crc kubenswrapper[4967]: I1008 00:07:45.628413 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:45 crc kubenswrapper[4967]: I1008 00:07:45.628429 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:45 crc kubenswrapper[4967]: I1008 00:07:45.628454 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:45 crc kubenswrapper[4967]: I1008 00:07:45.628470 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:45Z","lastTransitionTime":"2025-10-08T00:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:45 crc kubenswrapper[4967]: I1008 00:07:45.731302 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:45 crc kubenswrapper[4967]: I1008 00:07:45.731394 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:45 crc kubenswrapper[4967]: I1008 00:07:45.731420 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:45 crc kubenswrapper[4967]: I1008 00:07:45.731462 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:45 crc kubenswrapper[4967]: I1008 00:07:45.731487 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:45Z","lastTransitionTime":"2025-10-08T00:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:45 crc kubenswrapper[4967]: I1008 00:07:45.835219 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:45 crc kubenswrapper[4967]: I1008 00:07:45.835290 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:45 crc kubenswrapper[4967]: I1008 00:07:45.835309 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:45 crc kubenswrapper[4967]: I1008 00:07:45.835336 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:45 crc kubenswrapper[4967]: I1008 00:07:45.835357 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:45Z","lastTransitionTime":"2025-10-08T00:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:45 crc kubenswrapper[4967]: I1008 00:07:45.938904 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:45 crc kubenswrapper[4967]: I1008 00:07:45.938952 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:45 crc kubenswrapper[4967]: I1008 00:07:45.938970 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:45 crc kubenswrapper[4967]: I1008 00:07:45.938994 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:45 crc kubenswrapper[4967]: I1008 00:07:45.939013 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:45Z","lastTransitionTime":"2025-10-08T00:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:46 crc kubenswrapper[4967]: I1008 00:07:46.042287 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:46 crc kubenswrapper[4967]: I1008 00:07:46.042345 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:46 crc kubenswrapper[4967]: I1008 00:07:46.042366 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:46 crc kubenswrapper[4967]: I1008 00:07:46.042390 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:46 crc kubenswrapper[4967]: I1008 00:07:46.042407 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:46Z","lastTransitionTime":"2025-10-08T00:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:46 crc kubenswrapper[4967]: I1008 00:07:46.117912 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2h7z8" Oct 08 00:07:46 crc kubenswrapper[4967]: I1008 00:07:46.118017 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 00:07:46 crc kubenswrapper[4967]: E1008 00:07:46.118063 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2h7z8" podUID="59f68771-e25d-4367-87ee-68a31d767882" Oct 08 00:07:46 crc kubenswrapper[4967]: E1008 00:07:46.118201 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 00:07:46 crc kubenswrapper[4967]: I1008 00:07:46.145066 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:46 crc kubenswrapper[4967]: I1008 00:07:46.145119 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:46 crc kubenswrapper[4967]: I1008 00:07:46.145135 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:46 crc kubenswrapper[4967]: I1008 00:07:46.145156 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:46 crc kubenswrapper[4967]: I1008 00:07:46.145169 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:46Z","lastTransitionTime":"2025-10-08T00:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:46 crc kubenswrapper[4967]: I1008 00:07:46.248445 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:46 crc kubenswrapper[4967]: I1008 00:07:46.248507 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:46 crc kubenswrapper[4967]: I1008 00:07:46.248525 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:46 crc kubenswrapper[4967]: I1008 00:07:46.248549 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:46 crc kubenswrapper[4967]: I1008 00:07:46.248568 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:46Z","lastTransitionTime":"2025-10-08T00:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:46 crc kubenswrapper[4967]: I1008 00:07:46.351655 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:46 crc kubenswrapper[4967]: I1008 00:07:46.351720 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:46 crc kubenswrapper[4967]: I1008 00:07:46.351733 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:46 crc kubenswrapper[4967]: I1008 00:07:46.351753 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:46 crc kubenswrapper[4967]: I1008 00:07:46.351768 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:46Z","lastTransitionTime":"2025-10-08T00:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:46 crc kubenswrapper[4967]: I1008 00:07:46.454422 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:46 crc kubenswrapper[4967]: I1008 00:07:46.454476 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:46 crc kubenswrapper[4967]: I1008 00:07:46.454496 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:46 crc kubenswrapper[4967]: I1008 00:07:46.454515 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:46 crc kubenswrapper[4967]: I1008 00:07:46.454530 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:46Z","lastTransitionTime":"2025-10-08T00:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:46 crc kubenswrapper[4967]: I1008 00:07:46.556973 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:46 crc kubenswrapper[4967]: I1008 00:07:46.557021 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:46 crc kubenswrapper[4967]: I1008 00:07:46.557039 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:46 crc kubenswrapper[4967]: I1008 00:07:46.557062 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:46 crc kubenswrapper[4967]: I1008 00:07:46.557079 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:46Z","lastTransitionTime":"2025-10-08T00:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:46 crc kubenswrapper[4967]: I1008 00:07:46.660239 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:46 crc kubenswrapper[4967]: I1008 00:07:46.660336 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:46 crc kubenswrapper[4967]: I1008 00:07:46.660362 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:46 crc kubenswrapper[4967]: I1008 00:07:46.660392 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:46 crc kubenswrapper[4967]: I1008 00:07:46.660429 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:46Z","lastTransitionTime":"2025-10-08T00:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:46 crc kubenswrapper[4967]: I1008 00:07:46.763385 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:46 crc kubenswrapper[4967]: I1008 00:07:46.763439 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:46 crc kubenswrapper[4967]: I1008 00:07:46.763458 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:46 crc kubenswrapper[4967]: I1008 00:07:46.763479 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:46 crc kubenswrapper[4967]: I1008 00:07:46.763494 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:46Z","lastTransitionTime":"2025-10-08T00:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:46 crc kubenswrapper[4967]: I1008 00:07:46.866917 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:46 crc kubenswrapper[4967]: I1008 00:07:46.866981 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:46 crc kubenswrapper[4967]: I1008 00:07:46.866993 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:46 crc kubenswrapper[4967]: I1008 00:07:46.867016 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:46 crc kubenswrapper[4967]: I1008 00:07:46.867028 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:46Z","lastTransitionTime":"2025-10-08T00:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:46 crc kubenswrapper[4967]: I1008 00:07:46.970643 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:46 crc kubenswrapper[4967]: I1008 00:07:46.970711 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:46 crc kubenswrapper[4967]: I1008 00:07:46.970722 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:46 crc kubenswrapper[4967]: I1008 00:07:46.970743 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:46 crc kubenswrapper[4967]: I1008 00:07:46.970755 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:46Z","lastTransitionTime":"2025-10-08T00:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:47 crc kubenswrapper[4967]: I1008 00:07:47.072736 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:47 crc kubenswrapper[4967]: I1008 00:07:47.072788 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:47 crc kubenswrapper[4967]: I1008 00:07:47.072800 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:47 crc kubenswrapper[4967]: I1008 00:07:47.072818 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:47 crc kubenswrapper[4967]: I1008 00:07:47.072832 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:47Z","lastTransitionTime":"2025-10-08T00:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:47 crc kubenswrapper[4967]: I1008 00:07:47.116980 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 00:07:47 crc kubenswrapper[4967]: I1008 00:07:47.117030 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 00:07:47 crc kubenswrapper[4967]: E1008 00:07:47.117293 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 00:07:47 crc kubenswrapper[4967]: E1008 00:07:47.117478 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 00:07:47 crc kubenswrapper[4967]: I1008 00:07:47.131742 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Oct 08 00:07:47 crc kubenswrapper[4967]: I1008 00:07:47.175899 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:47 crc kubenswrapper[4967]: I1008 00:07:47.175939 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:47 crc kubenswrapper[4967]: I1008 00:07:47.175950 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:47 crc kubenswrapper[4967]: I1008 00:07:47.175968 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:47 crc kubenswrapper[4967]: I1008 00:07:47.175979 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:47Z","lastTransitionTime":"2025-10-08T00:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:47 crc kubenswrapper[4967]: I1008 00:07:47.278399 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:47 crc kubenswrapper[4967]: I1008 00:07:47.278478 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:47 crc kubenswrapper[4967]: I1008 00:07:47.278501 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:47 crc kubenswrapper[4967]: I1008 00:07:47.278530 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:47 crc kubenswrapper[4967]: I1008 00:07:47.278552 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:47Z","lastTransitionTime":"2025-10-08T00:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:47 crc kubenswrapper[4967]: I1008 00:07:47.380888 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:47 crc kubenswrapper[4967]: I1008 00:07:47.380953 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:47 crc kubenswrapper[4967]: I1008 00:07:47.380972 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:47 crc kubenswrapper[4967]: I1008 00:07:47.380999 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:47 crc kubenswrapper[4967]: I1008 00:07:47.381017 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:47Z","lastTransitionTime":"2025-10-08T00:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:47 crc kubenswrapper[4967]: I1008 00:07:47.483977 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:47 crc kubenswrapper[4967]: I1008 00:07:47.484015 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:47 crc kubenswrapper[4967]: I1008 00:07:47.484031 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:47 crc kubenswrapper[4967]: I1008 00:07:47.484052 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:47 crc kubenswrapper[4967]: I1008 00:07:47.484069 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:47Z","lastTransitionTime":"2025-10-08T00:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:47 crc kubenswrapper[4967]: I1008 00:07:47.587045 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:47 crc kubenswrapper[4967]: I1008 00:07:47.587345 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:47 crc kubenswrapper[4967]: I1008 00:07:47.587434 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:47 crc kubenswrapper[4967]: I1008 00:07:47.587582 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:47 crc kubenswrapper[4967]: I1008 00:07:47.587732 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:47Z","lastTransitionTime":"2025-10-08T00:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:47 crc kubenswrapper[4967]: I1008 00:07:47.690053 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:47 crc kubenswrapper[4967]: I1008 00:07:47.690352 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:47 crc kubenswrapper[4967]: I1008 00:07:47.690442 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:47 crc kubenswrapper[4967]: I1008 00:07:47.690560 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:47 crc kubenswrapper[4967]: I1008 00:07:47.690659 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:47Z","lastTransitionTime":"2025-10-08T00:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:47 crc kubenswrapper[4967]: I1008 00:07:47.794176 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:47 crc kubenswrapper[4967]: I1008 00:07:47.794231 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:47 crc kubenswrapper[4967]: I1008 00:07:47.794240 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:47 crc kubenswrapper[4967]: I1008 00:07:47.794255 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:47 crc kubenswrapper[4967]: I1008 00:07:47.794268 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:47Z","lastTransitionTime":"2025-10-08T00:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:47 crc kubenswrapper[4967]: I1008 00:07:47.896690 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:47 crc kubenswrapper[4967]: I1008 00:07:47.896718 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:47 crc kubenswrapper[4967]: I1008 00:07:47.896727 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:47 crc kubenswrapper[4967]: I1008 00:07:47.896740 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:47 crc kubenswrapper[4967]: I1008 00:07:47.896749 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:47Z","lastTransitionTime":"2025-10-08T00:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:47 crc kubenswrapper[4967]: I1008 00:07:47.999808 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:48 crc kubenswrapper[4967]: I1008 00:07:48.000098 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:48 crc kubenswrapper[4967]: I1008 00:07:48.000189 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:48 crc kubenswrapper[4967]: I1008 00:07:48.000299 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:48 crc kubenswrapper[4967]: I1008 00:07:48.000387 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:48Z","lastTransitionTime":"2025-10-08T00:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:48 crc kubenswrapper[4967]: I1008 00:07:48.102526 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:48 crc kubenswrapper[4967]: I1008 00:07:48.102806 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:48 crc kubenswrapper[4967]: I1008 00:07:48.102876 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:48 crc kubenswrapper[4967]: I1008 00:07:48.102938 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:48 crc kubenswrapper[4967]: I1008 00:07:48.103011 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:48Z","lastTransitionTime":"2025-10-08T00:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:48 crc kubenswrapper[4967]: I1008 00:07:48.117191 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2h7z8" Oct 08 00:07:48 crc kubenswrapper[4967]: E1008 00:07:48.117342 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2h7z8" podUID="59f68771-e25d-4367-87ee-68a31d767882" Oct 08 00:07:48 crc kubenswrapper[4967]: I1008 00:07:48.117230 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 00:07:48 crc kubenswrapper[4967]: E1008 00:07:48.117536 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 00:07:48 crc kubenswrapper[4967]: I1008 00:07:48.205243 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:48 crc kubenswrapper[4967]: I1008 00:07:48.205802 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:48 crc kubenswrapper[4967]: I1008 00:07:48.205872 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:48 crc kubenswrapper[4967]: I1008 00:07:48.205942 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:48 crc kubenswrapper[4967]: I1008 00:07:48.206000 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:48Z","lastTransitionTime":"2025-10-08T00:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:48 crc kubenswrapper[4967]: I1008 00:07:48.308216 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:48 crc kubenswrapper[4967]: I1008 00:07:48.308250 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:48 crc kubenswrapper[4967]: I1008 00:07:48.308260 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:48 crc kubenswrapper[4967]: I1008 00:07:48.308273 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:48 crc kubenswrapper[4967]: I1008 00:07:48.308281 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:48Z","lastTransitionTime":"2025-10-08T00:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:48 crc kubenswrapper[4967]: I1008 00:07:48.411229 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:48 crc kubenswrapper[4967]: I1008 00:07:48.411274 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:48 crc kubenswrapper[4967]: I1008 00:07:48.411286 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:48 crc kubenswrapper[4967]: I1008 00:07:48.411304 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:48 crc kubenswrapper[4967]: I1008 00:07:48.411316 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:48Z","lastTransitionTime":"2025-10-08T00:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:48 crc kubenswrapper[4967]: I1008 00:07:48.513786 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:48 crc kubenswrapper[4967]: I1008 00:07:48.513832 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:48 crc kubenswrapper[4967]: I1008 00:07:48.513849 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:48 crc kubenswrapper[4967]: I1008 00:07:48.513872 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:48 crc kubenswrapper[4967]: I1008 00:07:48.513890 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:48Z","lastTransitionTime":"2025-10-08T00:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:48 crc kubenswrapper[4967]: I1008 00:07:48.616444 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:48 crc kubenswrapper[4967]: I1008 00:07:48.616491 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:48 crc kubenswrapper[4967]: I1008 00:07:48.616504 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:48 crc kubenswrapper[4967]: I1008 00:07:48.616522 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:48 crc kubenswrapper[4967]: I1008 00:07:48.616535 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:48Z","lastTransitionTime":"2025-10-08T00:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:48 crc kubenswrapper[4967]: I1008 00:07:48.719154 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:48 crc kubenswrapper[4967]: I1008 00:07:48.719188 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:48 crc kubenswrapper[4967]: I1008 00:07:48.719197 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:48 crc kubenswrapper[4967]: I1008 00:07:48.719211 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:48 crc kubenswrapper[4967]: I1008 00:07:48.719220 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:48Z","lastTransitionTime":"2025-10-08T00:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:48 crc kubenswrapper[4967]: I1008 00:07:48.820792 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:48 crc kubenswrapper[4967]: I1008 00:07:48.820821 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:48 crc kubenswrapper[4967]: I1008 00:07:48.820832 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:48 crc kubenswrapper[4967]: I1008 00:07:48.820845 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:48 crc kubenswrapper[4967]: I1008 00:07:48.820853 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:48Z","lastTransitionTime":"2025-10-08T00:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:48 crc kubenswrapper[4967]: I1008 00:07:48.923302 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:48 crc kubenswrapper[4967]: I1008 00:07:48.923542 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:48 crc kubenswrapper[4967]: I1008 00:07:48.923614 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:48 crc kubenswrapper[4967]: I1008 00:07:48.923721 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:48 crc kubenswrapper[4967]: I1008 00:07:48.923808 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:48Z","lastTransitionTime":"2025-10-08T00:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:49 crc kubenswrapper[4967]: I1008 00:07:49.026269 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:49 crc kubenswrapper[4967]: I1008 00:07:49.026312 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:49 crc kubenswrapper[4967]: I1008 00:07:49.026324 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:49 crc kubenswrapper[4967]: I1008 00:07:49.026345 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:49 crc kubenswrapper[4967]: I1008 00:07:49.026357 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:49Z","lastTransitionTime":"2025-10-08T00:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:49 crc kubenswrapper[4967]: I1008 00:07:49.117823 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 00:07:49 crc kubenswrapper[4967]: E1008 00:07:49.117995 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 00:07:49 crc kubenswrapper[4967]: I1008 00:07:49.117850 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 00:07:49 crc kubenswrapper[4967]: E1008 00:07:49.118496 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 00:07:49 crc kubenswrapper[4967]: I1008 00:07:49.128758 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:49 crc kubenswrapper[4967]: I1008 00:07:49.130622 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:49 crc kubenswrapper[4967]: I1008 00:07:49.130790 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:49 crc kubenswrapper[4967]: I1008 00:07:49.130877 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:49 crc kubenswrapper[4967]: I1008 00:07:49.130967 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:49Z","lastTransitionTime":"2025-10-08T00:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:49 crc kubenswrapper[4967]: I1008 00:07:49.233283 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:49 crc kubenswrapper[4967]: I1008 00:07:49.233322 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:49 crc kubenswrapper[4967]: I1008 00:07:49.233334 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:49 crc kubenswrapper[4967]: I1008 00:07:49.233351 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:49 crc kubenswrapper[4967]: I1008 00:07:49.233362 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:49Z","lastTransitionTime":"2025-10-08T00:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:49 crc kubenswrapper[4967]: I1008 00:07:49.335771 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:49 crc kubenswrapper[4967]: I1008 00:07:49.335801 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:49 crc kubenswrapper[4967]: I1008 00:07:49.335812 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:49 crc kubenswrapper[4967]: I1008 00:07:49.335826 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:49 crc kubenswrapper[4967]: I1008 00:07:49.335836 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:49Z","lastTransitionTime":"2025-10-08T00:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:49 crc kubenswrapper[4967]: I1008 00:07:49.437618 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:49 crc kubenswrapper[4967]: I1008 00:07:49.437693 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:49 crc kubenswrapper[4967]: I1008 00:07:49.437707 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:49 crc kubenswrapper[4967]: I1008 00:07:49.437725 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:49 crc kubenswrapper[4967]: I1008 00:07:49.437736 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:49Z","lastTransitionTime":"2025-10-08T00:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:49 crc kubenswrapper[4967]: I1008 00:07:49.540051 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:49 crc kubenswrapper[4967]: I1008 00:07:49.540106 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:49 crc kubenswrapper[4967]: I1008 00:07:49.540119 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:49 crc kubenswrapper[4967]: I1008 00:07:49.540137 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:49 crc kubenswrapper[4967]: I1008 00:07:49.540149 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:49Z","lastTransitionTime":"2025-10-08T00:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:49 crc kubenswrapper[4967]: I1008 00:07:49.608480 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:49 crc kubenswrapper[4967]: I1008 00:07:49.608530 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:49 crc kubenswrapper[4967]: I1008 00:07:49.608539 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:49 crc kubenswrapper[4967]: I1008 00:07:49.608552 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:49 crc kubenswrapper[4967]: I1008 00:07:49.608560 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:49Z","lastTransitionTime":"2025-10-08T00:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:49 crc kubenswrapper[4967]: E1008 00:07:49.623889 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5cddad4c-22d1-42d7-a373-87b47c1cc97b\\\",\\\"systemUUID\\\":\\\"b243ca1c-20cd-4e6f-96b2-52fce74dd2d4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:49Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:49 crc kubenswrapper[4967]: I1008 00:07:49.628167 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:49 crc kubenswrapper[4967]: I1008 00:07:49.628205 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:49 crc kubenswrapper[4967]: I1008 00:07:49.628215 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:49 crc kubenswrapper[4967]: I1008 00:07:49.628229 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:49 crc kubenswrapper[4967]: I1008 00:07:49.628239 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:49Z","lastTransitionTime":"2025-10-08T00:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:49 crc kubenswrapper[4967]: E1008 00:07:49.643626 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5cddad4c-22d1-42d7-a373-87b47c1cc97b\\\",\\\"systemUUID\\\":\\\"b243ca1c-20cd-4e6f-96b2-52fce74dd2d4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:49Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:49 crc kubenswrapper[4967]: I1008 00:07:49.647963 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:49 crc kubenswrapper[4967]: I1008 00:07:49.648009 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:49 crc kubenswrapper[4967]: I1008 00:07:49.648022 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:49 crc kubenswrapper[4967]: I1008 00:07:49.648039 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:49 crc kubenswrapper[4967]: I1008 00:07:49.648050 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:49Z","lastTransitionTime":"2025-10-08T00:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:49 crc kubenswrapper[4967]: E1008 00:07:49.659584 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5cddad4c-22d1-42d7-a373-87b47c1cc97b\\\",\\\"systemUUID\\\":\\\"b243ca1c-20cd-4e6f-96b2-52fce74dd2d4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:49Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:49 crc kubenswrapper[4967]: I1008 00:07:49.663266 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:49 crc kubenswrapper[4967]: I1008 00:07:49.663309 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:49 crc kubenswrapper[4967]: I1008 00:07:49.663319 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:49 crc kubenswrapper[4967]: I1008 00:07:49.663334 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:49 crc kubenswrapper[4967]: I1008 00:07:49.663345 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:49Z","lastTransitionTime":"2025-10-08T00:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:49 crc kubenswrapper[4967]: E1008 00:07:49.676911 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5cddad4c-22d1-42d7-a373-87b47c1cc97b\\\",\\\"systemUUID\\\":\\\"b243ca1c-20cd-4e6f-96b2-52fce74dd2d4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:49Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:49 crc kubenswrapper[4967]: I1008 00:07:49.680542 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:49 crc kubenswrapper[4967]: I1008 00:07:49.680584 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:49 crc kubenswrapper[4967]: I1008 00:07:49.680595 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:49 crc kubenswrapper[4967]: I1008 00:07:49.680612 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:49 crc kubenswrapper[4967]: I1008 00:07:49.680627 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:49Z","lastTransitionTime":"2025-10-08T00:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:49 crc kubenswrapper[4967]: E1008 00:07:49.695144 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5cddad4c-22d1-42d7-a373-87b47c1cc97b\\\",\\\"systemUUID\\\":\\\"b243ca1c-20cd-4e6f-96b2-52fce74dd2d4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:49Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:49 crc kubenswrapper[4967]: E1008 00:07:49.695257 4967 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 08 00:07:49 crc kubenswrapper[4967]: I1008 00:07:49.696543 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:49 crc kubenswrapper[4967]: I1008 00:07:49.696592 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:49 crc kubenswrapper[4967]: I1008 00:07:49.696606 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:49 crc kubenswrapper[4967]: I1008 00:07:49.696624 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:49 crc kubenswrapper[4967]: I1008 00:07:49.696655 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:49Z","lastTransitionTime":"2025-10-08T00:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:49 crc kubenswrapper[4967]: I1008 00:07:49.798715 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:49 crc kubenswrapper[4967]: I1008 00:07:49.798755 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:49 crc kubenswrapper[4967]: I1008 00:07:49.798766 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:49 crc kubenswrapper[4967]: I1008 00:07:49.798782 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:49 crc kubenswrapper[4967]: I1008 00:07:49.798793 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:49Z","lastTransitionTime":"2025-10-08T00:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:49 crc kubenswrapper[4967]: I1008 00:07:49.900556 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:49 crc kubenswrapper[4967]: I1008 00:07:49.900595 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:49 crc kubenswrapper[4967]: I1008 00:07:49.900606 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:49 crc kubenswrapper[4967]: I1008 00:07:49.900620 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:49 crc kubenswrapper[4967]: I1008 00:07:49.900629 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:49Z","lastTransitionTime":"2025-10-08T00:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:50 crc kubenswrapper[4967]: I1008 00:07:50.003839 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:50 crc kubenswrapper[4967]: I1008 00:07:50.003889 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:50 crc kubenswrapper[4967]: I1008 00:07:50.003907 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:50 crc kubenswrapper[4967]: I1008 00:07:50.003929 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:50 crc kubenswrapper[4967]: I1008 00:07:50.003947 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:50Z","lastTransitionTime":"2025-10-08T00:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:50 crc kubenswrapper[4967]: I1008 00:07:50.106768 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:50 crc kubenswrapper[4967]: I1008 00:07:50.106821 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:50 crc kubenswrapper[4967]: I1008 00:07:50.106837 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:50 crc kubenswrapper[4967]: I1008 00:07:50.106857 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:50 crc kubenswrapper[4967]: I1008 00:07:50.106869 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:50Z","lastTransitionTime":"2025-10-08T00:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:50 crc kubenswrapper[4967]: I1008 00:07:50.117193 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 00:07:50 crc kubenswrapper[4967]: I1008 00:07:50.117307 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2h7z8" Oct 08 00:07:50 crc kubenswrapper[4967]: E1008 00:07:50.117316 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 00:07:50 crc kubenswrapper[4967]: E1008 00:07:50.117526 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2h7z8" podUID="59f68771-e25d-4367-87ee-68a31d767882" Oct 08 00:07:50 crc kubenswrapper[4967]: I1008 00:07:50.208877 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:50 crc kubenswrapper[4967]: I1008 00:07:50.209152 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:50 crc kubenswrapper[4967]: I1008 00:07:50.209249 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:50 crc kubenswrapper[4967]: I1008 00:07:50.209405 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:50 crc kubenswrapper[4967]: I1008 00:07:50.209602 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:50Z","lastTransitionTime":"2025-10-08T00:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:50 crc kubenswrapper[4967]: I1008 00:07:50.312327 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:50 crc kubenswrapper[4967]: I1008 00:07:50.312358 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:50 crc kubenswrapper[4967]: I1008 00:07:50.312369 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:50 crc kubenswrapper[4967]: I1008 00:07:50.312385 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:50 crc kubenswrapper[4967]: I1008 00:07:50.312396 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:50Z","lastTransitionTime":"2025-10-08T00:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:50 crc kubenswrapper[4967]: I1008 00:07:50.415978 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:50 crc kubenswrapper[4967]: I1008 00:07:50.416025 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:50 crc kubenswrapper[4967]: I1008 00:07:50.416045 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:50 crc kubenswrapper[4967]: I1008 00:07:50.416069 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:50 crc kubenswrapper[4967]: I1008 00:07:50.416087 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:50Z","lastTransitionTime":"2025-10-08T00:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:50 crc kubenswrapper[4967]: I1008 00:07:50.519044 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:50 crc kubenswrapper[4967]: I1008 00:07:50.519104 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:50 crc kubenswrapper[4967]: I1008 00:07:50.519122 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:50 crc kubenswrapper[4967]: I1008 00:07:50.519146 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:50 crc kubenswrapper[4967]: I1008 00:07:50.519163 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:50Z","lastTransitionTime":"2025-10-08T00:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:50 crc kubenswrapper[4967]: I1008 00:07:50.622296 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:50 crc kubenswrapper[4967]: I1008 00:07:50.622368 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:50 crc kubenswrapper[4967]: I1008 00:07:50.622388 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:50 crc kubenswrapper[4967]: I1008 00:07:50.622453 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:50 crc kubenswrapper[4967]: I1008 00:07:50.622471 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:50Z","lastTransitionTime":"2025-10-08T00:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:50 crc kubenswrapper[4967]: I1008 00:07:50.724450 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:50 crc kubenswrapper[4967]: I1008 00:07:50.724517 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:50 crc kubenswrapper[4967]: I1008 00:07:50.724537 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:50 crc kubenswrapper[4967]: I1008 00:07:50.724560 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:50 crc kubenswrapper[4967]: I1008 00:07:50.724575 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:50Z","lastTransitionTime":"2025-10-08T00:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:50 crc kubenswrapper[4967]: I1008 00:07:50.826753 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:50 crc kubenswrapper[4967]: I1008 00:07:50.826837 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:50 crc kubenswrapper[4967]: I1008 00:07:50.826854 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:50 crc kubenswrapper[4967]: I1008 00:07:50.826877 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:50 crc kubenswrapper[4967]: I1008 00:07:50.826893 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:50Z","lastTransitionTime":"2025-10-08T00:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:50 crc kubenswrapper[4967]: I1008 00:07:50.926730 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/59f68771-e25d-4367-87ee-68a31d767882-metrics-certs\") pod \"network-metrics-daemon-2h7z8\" (UID: \"59f68771-e25d-4367-87ee-68a31d767882\") " pod="openshift-multus/network-metrics-daemon-2h7z8" Oct 08 00:07:50 crc kubenswrapper[4967]: E1008 00:07:50.926932 4967 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 00:07:50 crc kubenswrapper[4967]: E1008 00:07:50.927166 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/59f68771-e25d-4367-87ee-68a31d767882-metrics-certs podName:59f68771-e25d-4367-87ee-68a31d767882 nodeName:}" failed. No retries permitted until 2025-10-08 00:08:22.927148942 +0000 UTC m=+101.577683601 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/59f68771-e25d-4367-87ee-68a31d767882-metrics-certs") pod "network-metrics-daemon-2h7z8" (UID: "59f68771-e25d-4367-87ee-68a31d767882") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 00:07:50 crc kubenswrapper[4967]: I1008 00:07:50.929183 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:50 crc kubenswrapper[4967]: I1008 00:07:50.929322 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:50 crc kubenswrapper[4967]: I1008 00:07:50.929442 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:50 crc kubenswrapper[4967]: I1008 00:07:50.929576 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:50 crc kubenswrapper[4967]: I1008 00:07:50.929717 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:50Z","lastTransitionTime":"2025-10-08T00:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:51 crc kubenswrapper[4967]: I1008 00:07:51.033268 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:51 crc kubenswrapper[4967]: I1008 00:07:51.033335 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:51 crc kubenswrapper[4967]: I1008 00:07:51.033352 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:51 crc kubenswrapper[4967]: I1008 00:07:51.033379 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:51 crc kubenswrapper[4967]: I1008 00:07:51.033396 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:51Z","lastTransitionTime":"2025-10-08T00:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:51 crc kubenswrapper[4967]: I1008 00:07:51.116857 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 00:07:51 crc kubenswrapper[4967]: E1008 00:07:51.116991 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 00:07:51 crc kubenswrapper[4967]: I1008 00:07:51.117306 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 00:07:51 crc kubenswrapper[4967]: E1008 00:07:51.117614 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 00:07:51 crc kubenswrapper[4967]: I1008 00:07:51.136417 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:51 crc kubenswrapper[4967]: I1008 00:07:51.136464 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:51 crc kubenswrapper[4967]: I1008 00:07:51.136478 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:51 crc kubenswrapper[4967]: I1008 00:07:51.136512 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:51 crc kubenswrapper[4967]: I1008 00:07:51.136526 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:51Z","lastTransitionTime":"2025-10-08T00:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:51 crc kubenswrapper[4967]: I1008 00:07:51.239256 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:51 crc kubenswrapper[4967]: I1008 00:07:51.239331 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:51 crc kubenswrapper[4967]: I1008 00:07:51.239356 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:51 crc kubenswrapper[4967]: I1008 00:07:51.239386 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:51 crc kubenswrapper[4967]: I1008 00:07:51.239407 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:51Z","lastTransitionTime":"2025-10-08T00:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:51 crc kubenswrapper[4967]: I1008 00:07:51.342917 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:51 crc kubenswrapper[4967]: I1008 00:07:51.342968 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:51 crc kubenswrapper[4967]: I1008 00:07:51.342987 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:51 crc kubenswrapper[4967]: I1008 00:07:51.343012 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:51 crc kubenswrapper[4967]: I1008 00:07:51.343031 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:51Z","lastTransitionTime":"2025-10-08T00:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:51 crc kubenswrapper[4967]: I1008 00:07:51.446042 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:51 crc kubenswrapper[4967]: I1008 00:07:51.446123 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:51 crc kubenswrapper[4967]: I1008 00:07:51.446149 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:51 crc kubenswrapper[4967]: I1008 00:07:51.446180 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:51 crc kubenswrapper[4967]: I1008 00:07:51.446204 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:51Z","lastTransitionTime":"2025-10-08T00:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:51 crc kubenswrapper[4967]: I1008 00:07:51.548396 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:51 crc kubenswrapper[4967]: I1008 00:07:51.548423 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:51 crc kubenswrapper[4967]: I1008 00:07:51.548434 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:51 crc kubenswrapper[4967]: I1008 00:07:51.548449 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:51 crc kubenswrapper[4967]: I1008 00:07:51.548458 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:51Z","lastTransitionTime":"2025-10-08T00:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:51 crc kubenswrapper[4967]: I1008 00:07:51.650523 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:51 crc kubenswrapper[4967]: I1008 00:07:51.650868 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:51 crc kubenswrapper[4967]: I1008 00:07:51.650945 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:51 crc kubenswrapper[4967]: I1008 00:07:51.651024 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:51 crc kubenswrapper[4967]: I1008 00:07:51.651089 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:51Z","lastTransitionTime":"2025-10-08T00:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:51 crc kubenswrapper[4967]: I1008 00:07:51.754912 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:51 crc kubenswrapper[4967]: I1008 00:07:51.755206 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:51 crc kubenswrapper[4967]: I1008 00:07:51.755316 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:51 crc kubenswrapper[4967]: I1008 00:07:51.755630 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:51 crc kubenswrapper[4967]: I1008 00:07:51.755773 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:51Z","lastTransitionTime":"2025-10-08T00:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:51 crc kubenswrapper[4967]: I1008 00:07:51.859122 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:51 crc kubenswrapper[4967]: I1008 00:07:51.859174 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:51 crc kubenswrapper[4967]: I1008 00:07:51.859191 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:51 crc kubenswrapper[4967]: I1008 00:07:51.859213 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:51 crc kubenswrapper[4967]: I1008 00:07:51.859228 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:51Z","lastTransitionTime":"2025-10-08T00:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:51 crc kubenswrapper[4967]: I1008 00:07:51.961696 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:51 crc kubenswrapper[4967]: I1008 00:07:51.961745 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:51 crc kubenswrapper[4967]: I1008 00:07:51.961758 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:51 crc kubenswrapper[4967]: I1008 00:07:51.961776 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:51 crc kubenswrapper[4967]: I1008 00:07:51.961788 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:51Z","lastTransitionTime":"2025-10-08T00:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.065042 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.065082 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.065094 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.065111 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.065123 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:52Z","lastTransitionTime":"2025-10-08T00:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.117649 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2h7z8" Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.117695 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 00:07:52 crc kubenswrapper[4967]: E1008 00:07:52.117782 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2h7z8" podUID="59f68771-e25d-4367-87ee-68a31d767882" Oct 08 00:07:52 crc kubenswrapper[4967]: E1008 00:07:52.117831 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.132288 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:52Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.149380 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-npz2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c71501b1-aa9e-420e-b1bd-7e15365e7adc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dfe6c420d804b2d7daef4b5cfe3f87b77b369287c4d65553fa93f5f24f589d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gblk2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-npz2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:52Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.161756 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2h7z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59f68771-e25d-4367-87ee-68a31d767882\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7r2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7r2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:18Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2h7z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:52Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.167185 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.167211 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.167220 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.167234 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.167244 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:52Z","lastTransitionTime":"2025-10-08T00:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.172500 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6ea2440-b2ae-481e-9442-0b3dda9688ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37fd5dcdfb864320d4d17729ab48c6da505fd867f384ad70a6dcf5bed78a3978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c7893dfaa0370f159d41c30cd869bb18550a819e8c3d20dd6f57dcbcc697a57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c430979d298477b4394c3567bb3800ab7dea874cee8fa41a6a50a19a7e3a0133\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://426e86ab409776ff0fa9505dc0d8ec36f27a5bf21d3aedafb19d3935bdba7529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://426e86ab409776ff0fa9505dc0d8ec36f27a5bf21d3aedafb19d3935bdba7529\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:52Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.184973 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58c4704-a88a-43b4-b944-4de626f13f7e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22f5dabb39da0a7bd01db4dae1437b9a231a1fcc0dfe749a4a5a637a72cf33c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01a2b3dc785520c79eae998482e5a1347725c76584357073238fe60a619edb77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6aa6f5e81feceab3d1c7587152a70546b365e629a69267e4d3084324a1de764a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63c2b7c4b2791bd1f16888c58b878dbff1f87884e68c9949de914d0d3e1d39fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:52Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.197869 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:52Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.210478 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9mrr7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9299bc4f-f8d3-4baf-9885-39573982fd64\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://994beafece9dead0d85ae74dc453a142460705fd66b52ba22f2fa799cf27da08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb49c8476b482416588c4d403016940730064a619c0641eb029c947a4bab1215\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb49c8476b482416588c4d403016940730064a619c0641eb029c947a4bab1215\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b7b63962dcbf477f1fa4d71826a4aa3358920a2c1b2f5e48716411b5bdca049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b7b63962dcbf477f1fa4d71826a4aa3358920a2c1b2f5e48716411b5bdca049\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1e770dc737173fd4f3220ed6c52b422ccbd829722386a37029e1c3e10d80e1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1e770dc737173fd4f3220ed6c52b422ccbd829722386a37029e1c3e10d80e1a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e357d970878c996f219a79f1876b363176c9847b9575f33090f62bc299529a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e357d970878c996f219a79f1876b363176c9847b9575f33090f62bc299529a95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f098966bfad49d06278eaa72be042596d488115e597737d9a0a0cfc78f99f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49f098966bfad49d06278eaa72be042596d488115e597737d9a0a0cfc78f99f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1d53876b7b863c612d597fd716c3260e8857b6542802073bb730f6756071b46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c1d53876b7b863c612d597fd716c3260e8857b6542802073bb730f6756071b46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9mrr7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:52Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.221057 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kpmlg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f975d47d-42a6-422e-80e2-d1ee282817fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a317ae6640eb38b3aee0f9738cca890e8ab3d92075b620cd312f40a0628d46c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qznqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ef61854aea40924c5372b0553cd0b7e8faa39a1d42fec5b65cf867c368f72c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qznqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-kpmlg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:52Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.234514 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc98bc5e-fbae-4c67-a5c3-9738b3e966cd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acfae51bf8bee1b379a3859b7fb98ad0df3bed5d49dc0abe82b917512955c009\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae14677731c53e87871129d2140a34b18f0a48555309b652ebd490996fb1a723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae14677731c53e87871129d2140a34b18f0a48555309b652ebd490996fb1a723\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:52Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.245241 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:52Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.257799 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbfcdc882e2be1cb9cf666c9e7775671078f7dbbfe8421dd15db737b89920a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:52Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.270057 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.270099 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.270111 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.270127 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.270139 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:52Z","lastTransitionTime":"2025-10-08T00:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.270568 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9461162a-c27d-465e-9ca7-2d3936877cd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9778d78ddc7325abd4d77bc5e71b0b114f50be5dda4a438c7cb6b05640c8a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z9tp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28760af04ef321d0f91b2bce3ad074d44b63dabf676cc37f3be6aa8726341a2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z9tp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xvdtp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:52Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.285505 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fa4b747-ab6e-4a11-9903-ef2dcb9668c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbadcdade274b6d9be940f22bc949e3c55cadf95dcad43c8cb7a1085c0e6708c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d0a0b2ef4a98ee828688081a74f4b7ecc31c8fae415d7257a560ecd43960ce3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96b53b1b283025b3d078a3f505d469d1872ad4bb8ab8e612c36c70b0b00a920\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef86a2422616e1e0d5c3afe844818c2bdf2150f941d1aefc506995f58587b2bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1400e110997c64fa33d6b8672c7e91cab5f782952ddf8e1dea2a882689026e7f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 00:06:56.760088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 00:06:56.761210 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-369805205/tls.crt::/tmp/serving-cert-369805205/tls.key\\\\\\\"\\\\nI1008 00:07:02.321448 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 00:07:02.325932 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 00:07:02.325980 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 00:07:02.326025 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 00:07:02.326042 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 00:07:02.336203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 00:07:02.336240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 00:07:02.336245 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 00:07:02.336250 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 00:07:02.336253 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 00:07:02.336256 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 00:07:02.336258 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 00:07:02.336430 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 00:07:02.340774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:46Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca28d8db19036e47a7ef3dadb70941f6c6cf5735e7703977d036185c46fce40\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5b222ad84c31d8746dfe03124c7287b401deb3ca2c62ce2b4fbd40e638cee0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5b222ad84c31d8746dfe03124c7287b401deb3ca2c62ce2b4fbd40e638cee0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:52Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.299697 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e98666fd3152bab57d3723e7da72aae2703b79f7ca73eb3c3e5f0e062a82592\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cf1cbe5548146fe359aafa351864125ec6b8bdd7b796aa4d311f6f3a16ad212\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:52Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.312136 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9tdbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41264a12-9486-4fee-a08c-607b590e5a70\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://806e85df796ae300d279f73a0cc4433a5bb1f7620ee8ce33c8dc4674c4ce0596\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5blv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9tdbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:52Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.340865 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f1af2ba-8907-4d14-afa1-ffde207c80f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4535cffb8c9d64e662e282ecb44f6817b4604866f0bb75015f3b888d5554f821\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b03b83ddff429dbfa8fe052901aaa86a39ff0dd75477ffa8a6b6f4198f5819ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c9fa0559eb35ba67e11e36069ac8d0b2fec76399788eef668a08ad4353df3ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3312263e12bd42fbed0476ab5ace66b5497f62c9477865e130d67e3922e05cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e69dc603055aa221d44d11c049ddcc50cec568979911a310ed773a90341a499\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115f7a33f3dedae8f59d2f45f465f95f48b99c824e895ef7a0be9fe7a1b61d23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1cfd26b02d0e17b111b8ad84dab6c2adf692de66ff77158ee7cbff941c66ecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1cfd26b02d0e17b111b8ad84dab6c2adf692de66ff77158ee7cbff941c66ecd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T00:07:29Z\\\",\\\"message\\\":\\\"iled objects of type *v1.Pod\\\\nI1008 00:07:28.973629 6638 obj_retry.go:409] Going to retry *v1.Pod resource setup for 1 objects: [openshift-multus/network-metrics-daemon-2h7z8]\\\\nI1008 00:07:28.973646 6638 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1008 00:07:28.973685 6638 obj_retry.go:285] Attempting retry of *v1.Pod openshift-multus/network-metrics-daemon-2h7z8 before timer (time: 2025-10-08 00:07:30.377116097 +0000 UTC m=+1.948919645): skip\\\\nI1008 00:07:28.973706 6638 obj_retry.go:420] Function iterateRetryResources for *v1.Pod ended (in 79.172µs)\\\\nI1008 00:07:28.973904 6638 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1008 00:07:28.973958 6638 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1008 00:07:28.974084 6638 factory.go:656] Stopping watch factory\\\\nI1008 00:07:28.974107 6638 ovnkube.go:599] Stopped ovnkube\\\\nI1008 00:07:28.974137 6638 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1008 00:07:28.974149 6638 handler.go:208] Removed *v1.Node event handler 2\\\\nI1008 00:07:28.974163 6638 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1008 00:07:28.974267 6638 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-bs8w9_openshift-ovn-kubernetes(4f1af2ba-8907-4d14-afa1-ffde207c80f4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27fb2b1225e386dcd7fecaaa0aa54ba1fdfafabfde6c3824ae07315313067542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bs8w9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:52Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.352192 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-v5n7t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abe7ae2d-7e63-4249-96e1-4000e9628ccc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52575e853ff6727dde7c89e4add4362be126127db27191c2a7bf2871efa719d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpsjd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:06Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-v5n7t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:52Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.368618 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94e63bc5641ce80d507f76a19b203f78c42368a7a4105f3fd09e19af21a1b735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:52Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.372213 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.372242 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.372254 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.372270 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.372282 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:52Z","lastTransitionTime":"2025-10-08T00:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.474239 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.474290 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.474303 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.474321 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.474332 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:52Z","lastTransitionTime":"2025-10-08T00:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.576362 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.576398 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.576409 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.576426 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.576437 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:52Z","lastTransitionTime":"2025-10-08T00:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.617314 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-npz2s_c71501b1-aa9e-420e-b1bd-7e15365e7adc/kube-multus/0.log" Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.617359 4967 generic.go:334] "Generic (PLEG): container finished" podID="c71501b1-aa9e-420e-b1bd-7e15365e7adc" containerID="82dfe6c420d804b2d7daef4b5cfe3f87b77b369287c4d65553fa93f5f24f589d" exitCode=1 Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.617386 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-npz2s" event={"ID":"c71501b1-aa9e-420e-b1bd-7e15365e7adc","Type":"ContainerDied","Data":"82dfe6c420d804b2d7daef4b5cfe3f87b77b369287c4d65553fa93f5f24f589d"} Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.617721 4967 scope.go:117] "RemoveContainer" containerID="82dfe6c420d804b2d7daef4b5cfe3f87b77b369287c4d65553fa93f5f24f589d" Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.632831 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58c4704-a88a-43b4-b944-4de626f13f7e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22f5dabb39da0a7bd01db4dae1437b9a231a1fcc0dfe749a4a5a637a72cf33c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01a2b3dc785520c79eae998482e5a1347725c76584357073238fe60a619edb77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6aa6f5e81feceab3d1c7587152a70546b365e629a69267e4d3084324a1de764a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63c2b7c4b2791bd1f16888c58b878dbff1f87884e68c9949de914d0d3e1d39fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:52Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.643895 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:52Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.658581 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9mrr7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9299bc4f-f8d3-4baf-9885-39573982fd64\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://994beafece9dead0d85ae74dc453a142460705fd66b52ba22f2fa799cf27da08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb49c8476b482416588c4d403016940730064a619c0641eb029c947a4bab1215\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb49c8476b482416588c4d403016940730064a619c0641eb029c947a4bab1215\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b7b63962dcbf477f1fa4d71826a4aa3358920a2c1b2f5e48716411b5bdca049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b7b63962dcbf477f1fa4d71826a4aa3358920a2c1b2f5e48716411b5bdca049\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1e770dc737173fd4f3220ed6c52b422ccbd829722386a37029e1c3e10d80e1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1e770dc737173fd4f3220ed6c52b422ccbd829722386a37029e1c3e10d80e1a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e357d970878c996f219a79f1876b363176c9847b9575f33090f62bc299529a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e357d970878c996f219a79f1876b363176c9847b9575f33090f62bc299529a95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f098966bfad49d06278eaa72be042596d488115e597737d9a0a0cfc78f99f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49f098966bfad49d06278eaa72be042596d488115e597737d9a0a0cfc78f99f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1d53876b7b863c612d597fd716c3260e8857b6542802073bb730f6756071b46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c1d53876b7b863c612d597fd716c3260e8857b6542802073bb730f6756071b46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9mrr7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:52Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.669140 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kpmlg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f975d47d-42a6-422e-80e2-d1ee282817fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a317ae6640eb38b3aee0f9738cca890e8ab3d92075b620cd312f40a0628d46c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qznqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ef61854aea40924c5372b0553cd0b7e8faa39a1d42fec5b65cf867c368f72c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qznqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-kpmlg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:52Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.678619 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.678660 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.678686 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.678701 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.678710 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:52Z","lastTransitionTime":"2025-10-08T00:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.681206 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc98bc5e-fbae-4c67-a5c3-9738b3e966cd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acfae51bf8bee1b379a3859b7fb98ad0df3bed5d49dc0abe82b917512955c009\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae14677731c53e87871129d2140a34b18f0a48555309b652ebd490996fb1a723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae14677731c53e87871129d2140a34b18f0a48555309b652ebd490996fb1a723\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:52Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.692502 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:52Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.705212 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbfcdc882e2be1cb9cf666c9e7775671078f7dbbfe8421dd15db737b89920a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:52Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.718929 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9461162a-c27d-465e-9ca7-2d3936877cd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9778d78ddc7325abd4d77bc5e71b0b114f50be5dda4a438c7cb6b05640c8a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z9tp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28760af04ef321d0f91b2bce3ad074d44b63dabf676cc37f3be6aa8726341a2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z9tp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xvdtp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:52Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.733593 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fa4b747-ab6e-4a11-9903-ef2dcb9668c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbadcdade274b6d9be940f22bc949e3c55cadf95dcad43c8cb7a1085c0e6708c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d0a0b2ef4a98ee828688081a74f4b7ecc31c8fae415d7257a560ecd43960ce3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96b53b1b283025b3d078a3f505d469d1872ad4bb8ab8e612c36c70b0b00a920\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef86a2422616e1e0d5c3afe844818c2bdf2150f941d1aefc506995f58587b2bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1400e110997c64fa33d6b8672c7e91cab5f782952ddf8e1dea2a882689026e7f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 00:06:56.760088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 00:06:56.761210 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-369805205/tls.crt::/tmp/serving-cert-369805205/tls.key\\\\\\\"\\\\nI1008 00:07:02.321448 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 00:07:02.325932 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 00:07:02.325980 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 00:07:02.326025 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 00:07:02.326042 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 00:07:02.336203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 00:07:02.336240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 00:07:02.336245 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 00:07:02.336250 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 00:07:02.336253 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 00:07:02.336256 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 00:07:02.336258 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 00:07:02.336430 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 00:07:02.340774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:46Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca28d8db19036e47a7ef3dadb70941f6c6cf5735e7703977d036185c46fce40\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5b222ad84c31d8746dfe03124c7287b401deb3ca2c62ce2b4fbd40e638cee0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5b222ad84c31d8746dfe03124c7287b401deb3ca2c62ce2b4fbd40e638cee0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:52Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.745365 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e98666fd3152bab57d3723e7da72aae2703b79f7ca73eb3c3e5f0e062a82592\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cf1cbe5548146fe359aafa351864125ec6b8bdd7b796aa4d311f6f3a16ad212\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:52Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.755715 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9tdbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41264a12-9486-4fee-a08c-607b590e5a70\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://806e85df796ae300d279f73a0cc4433a5bb1f7620ee8ce33c8dc4674c4ce0596\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5blv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9tdbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:52Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.771373 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f1af2ba-8907-4d14-afa1-ffde207c80f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4535cffb8c9d64e662e282ecb44f6817b4604866f0bb75015f3b888d5554f821\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b03b83ddff429dbfa8fe052901aaa86a39ff0dd75477ffa8a6b6f4198f5819ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c9fa0559eb35ba67e11e36069ac8d0b2fec76399788eef668a08ad4353df3ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3312263e12bd42fbed0476ab5ace66b5497f62c9477865e130d67e3922e05cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e69dc603055aa221d44d11c049ddcc50cec568979911a310ed773a90341a499\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115f7a33f3dedae8f59d2f45f465f95f48b99c824e895ef7a0be9fe7a1b61d23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1cfd26b02d0e17b111b8ad84dab6c2adf692de66ff77158ee7cbff941c66ecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1cfd26b02d0e17b111b8ad84dab6c2adf692de66ff77158ee7cbff941c66ecd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T00:07:29Z\\\",\\\"message\\\":\\\"iled objects of type *v1.Pod\\\\nI1008 00:07:28.973629 6638 obj_retry.go:409] Going to retry *v1.Pod resource setup for 1 objects: [openshift-multus/network-metrics-daemon-2h7z8]\\\\nI1008 00:07:28.973646 6638 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1008 00:07:28.973685 6638 obj_retry.go:285] Attempting retry of *v1.Pod openshift-multus/network-metrics-daemon-2h7z8 before timer (time: 2025-10-08 00:07:30.377116097 +0000 UTC m=+1.948919645): skip\\\\nI1008 00:07:28.973706 6638 obj_retry.go:420] Function iterateRetryResources for *v1.Pod ended (in 79.172µs)\\\\nI1008 00:07:28.973904 6638 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1008 00:07:28.973958 6638 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1008 00:07:28.974084 6638 factory.go:656] Stopping watch factory\\\\nI1008 00:07:28.974107 6638 ovnkube.go:599] Stopped ovnkube\\\\nI1008 00:07:28.974137 6638 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1008 00:07:28.974149 6638 handler.go:208] Removed *v1.Node event handler 2\\\\nI1008 00:07:28.974163 6638 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1008 00:07:28.974267 6638 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-bs8w9_openshift-ovn-kubernetes(4f1af2ba-8907-4d14-afa1-ffde207c80f4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27fb2b1225e386dcd7fecaaa0aa54ba1fdfafabfde6c3824ae07315313067542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bs8w9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:52Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.780737 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.780765 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.780774 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.780786 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.780794 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:52Z","lastTransitionTime":"2025-10-08T00:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.782575 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-v5n7t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abe7ae2d-7e63-4249-96e1-4000e9628ccc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52575e853ff6727dde7c89e4add4362be126127db27191c2a7bf2871efa719d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpsjd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:06Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-v5n7t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:52Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.795768 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94e63bc5641ce80d507f76a19b203f78c42368a7a4105f3fd09e19af21a1b735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:52Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.807824 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:52Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.819889 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-npz2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c71501b1-aa9e-420e-b1bd-7e15365e7adc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:52Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://82dfe6c420d804b2d7daef4b5cfe3f87b77b369287c4d65553fa93f5f24f589d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82dfe6c420d804b2d7daef4b5cfe3f87b77b369287c4d65553fa93f5f24f589d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T00:07:51Z\\\",\\\"message\\\":\\\"2025-10-08T00:07:06+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_b0e5397b-daec-448a-8e92-e00fdfae8aea\\\\n2025-10-08T00:07:06+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_b0e5397b-daec-448a-8e92-e00fdfae8aea to /host/opt/cni/bin/\\\\n2025-10-08T00:07:06Z [verbose] multus-daemon started\\\\n2025-10-08T00:07:06Z [verbose] Readiness Indicator file check\\\\n2025-10-08T00:07:51Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gblk2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-npz2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:52Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.837488 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2h7z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59f68771-e25d-4367-87ee-68a31d767882\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7r2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7r2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:18Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2h7z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:52Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.850084 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6ea2440-b2ae-481e-9442-0b3dda9688ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37fd5dcdfb864320d4d17729ab48c6da505fd867f384ad70a6dcf5bed78a3978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c7893dfaa0370f159d41c30cd869bb18550a819e8c3d20dd6f57dcbcc697a57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c430979d298477b4394c3567bb3800ab7dea874cee8fa41a6a50a19a7e3a0133\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://426e86ab409776ff0fa9505dc0d8ec36f27a5bf21d3aedafb19d3935bdba7529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://426e86ab409776ff0fa9505dc0d8ec36f27a5bf21d3aedafb19d3935bdba7529\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:52Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.882775 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.882815 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.882828 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.882844 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.882855 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:52Z","lastTransitionTime":"2025-10-08T00:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.985754 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.985797 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.985811 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.985827 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:52 crc kubenswrapper[4967]: I1008 00:07:52.985837 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:52Z","lastTransitionTime":"2025-10-08T00:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:53 crc kubenswrapper[4967]: I1008 00:07:53.088371 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:53 crc kubenswrapper[4967]: I1008 00:07:53.088415 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:53 crc kubenswrapper[4967]: I1008 00:07:53.088426 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:53 crc kubenswrapper[4967]: I1008 00:07:53.088442 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:53 crc kubenswrapper[4967]: I1008 00:07:53.088451 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:53Z","lastTransitionTime":"2025-10-08T00:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:53 crc kubenswrapper[4967]: I1008 00:07:53.117753 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 00:07:53 crc kubenswrapper[4967]: I1008 00:07:53.117824 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 00:07:53 crc kubenswrapper[4967]: E1008 00:07:53.117871 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 00:07:53 crc kubenswrapper[4967]: E1008 00:07:53.117941 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 00:07:53 crc kubenswrapper[4967]: I1008 00:07:53.190609 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:53 crc kubenswrapper[4967]: I1008 00:07:53.190642 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:53 crc kubenswrapper[4967]: I1008 00:07:53.190652 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:53 crc kubenswrapper[4967]: I1008 00:07:53.190680 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:53 crc kubenswrapper[4967]: I1008 00:07:53.190689 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:53Z","lastTransitionTime":"2025-10-08T00:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:53 crc kubenswrapper[4967]: I1008 00:07:53.293135 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:53 crc kubenswrapper[4967]: I1008 00:07:53.293174 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:53 crc kubenswrapper[4967]: I1008 00:07:53.293183 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:53 crc kubenswrapper[4967]: I1008 00:07:53.293197 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:53 crc kubenswrapper[4967]: I1008 00:07:53.293206 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:53Z","lastTransitionTime":"2025-10-08T00:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:53 crc kubenswrapper[4967]: I1008 00:07:53.395613 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:53 crc kubenswrapper[4967]: I1008 00:07:53.395658 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:53 crc kubenswrapper[4967]: I1008 00:07:53.395691 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:53 crc kubenswrapper[4967]: I1008 00:07:53.395708 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:53 crc kubenswrapper[4967]: I1008 00:07:53.395720 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:53Z","lastTransitionTime":"2025-10-08T00:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:53 crc kubenswrapper[4967]: I1008 00:07:53.502959 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:53 crc kubenswrapper[4967]: I1008 00:07:53.502993 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:53 crc kubenswrapper[4967]: I1008 00:07:53.503001 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:53 crc kubenswrapper[4967]: I1008 00:07:53.503015 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:53 crc kubenswrapper[4967]: I1008 00:07:53.503024 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:53Z","lastTransitionTime":"2025-10-08T00:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:53 crc kubenswrapper[4967]: I1008 00:07:53.605383 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:53 crc kubenswrapper[4967]: I1008 00:07:53.605448 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:53 crc kubenswrapper[4967]: I1008 00:07:53.605465 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:53 crc kubenswrapper[4967]: I1008 00:07:53.605486 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:53 crc kubenswrapper[4967]: I1008 00:07:53.605498 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:53Z","lastTransitionTime":"2025-10-08T00:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:53 crc kubenswrapper[4967]: I1008 00:07:53.621331 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-npz2s_c71501b1-aa9e-420e-b1bd-7e15365e7adc/kube-multus/0.log" Oct 08 00:07:53 crc kubenswrapper[4967]: I1008 00:07:53.621527 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-npz2s" event={"ID":"c71501b1-aa9e-420e-b1bd-7e15365e7adc","Type":"ContainerStarted","Data":"9010a5d6165a64f986f9101dbbbcaf8a7d3fa5af85c14be4bf3aa8ed8536ad4e"} Oct 08 00:07:53 crc kubenswrapper[4967]: I1008 00:07:53.638913 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9mrr7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9299bc4f-f8d3-4baf-9885-39573982fd64\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://994beafece9dead0d85ae74dc453a142460705fd66b52ba22f2fa799cf27da08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb49c8476b482416588c4d403016940730064a619c0641eb029c947a4bab1215\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb49c8476b482416588c4d403016940730064a619c0641eb029c947a4bab1215\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b7b63962dcbf477f1fa4d71826a4aa3358920a2c1b2f5e48716411b5bdca049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b7b63962dcbf477f1fa4d71826a4aa3358920a2c1b2f5e48716411b5bdca049\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1e770dc737173fd4f3220ed6c52b422ccbd829722386a37029e1c3e10d80e1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1e770dc737173fd4f3220ed6c52b422ccbd829722386a37029e1c3e10d80e1a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e357d970878c996f219a79f1876b363176c9847b9575f33090f62bc299529a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e357d970878c996f219a79f1876b363176c9847b9575f33090f62bc299529a95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f098966bfad49d06278eaa72be042596d488115e597737d9a0a0cfc78f99f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49f098966bfad49d06278eaa72be042596d488115e597737d9a0a0cfc78f99f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1d53876b7b863c612d597fd716c3260e8857b6542802073bb730f6756071b46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c1d53876b7b863c612d597fd716c3260e8857b6542802073bb730f6756071b46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9mrr7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:53Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:53 crc kubenswrapper[4967]: I1008 00:07:53.651466 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kpmlg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f975d47d-42a6-422e-80e2-d1ee282817fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a317ae6640eb38b3aee0f9738cca890e8ab3d92075b620cd312f40a0628d46c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qznqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ef61854aea40924c5372b0553cd0b7e8faa39a1d42fec5b65cf867c368f72c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qznqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-kpmlg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:53Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:53 crc kubenswrapper[4967]: I1008 00:07:53.664154 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58c4704-a88a-43b4-b944-4de626f13f7e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22f5dabb39da0a7bd01db4dae1437b9a231a1fcc0dfe749a4a5a637a72cf33c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01a2b3dc785520c79eae998482e5a1347725c76584357073238fe60a619edb77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6aa6f5e81feceab3d1c7587152a70546b365e629a69267e4d3084324a1de764a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63c2b7c4b2791bd1f16888c58b878dbff1f87884e68c9949de914d0d3e1d39fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:53Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:53 crc kubenswrapper[4967]: I1008 00:07:53.675046 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:53Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:53 crc kubenswrapper[4967]: I1008 00:07:53.687327 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbfcdc882e2be1cb9cf666c9e7775671078f7dbbfe8421dd15db737b89920a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:53Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:53 crc kubenswrapper[4967]: I1008 00:07:53.698395 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9461162a-c27d-465e-9ca7-2d3936877cd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9778d78ddc7325abd4d77bc5e71b0b114f50be5dda4a438c7cb6b05640c8a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z9tp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28760af04ef321d0f91b2bce3ad074d44b63dabf676cc37f3be6aa8726341a2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z9tp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xvdtp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:53Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:53 crc kubenswrapper[4967]: I1008 00:07:53.708020 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:53 crc kubenswrapper[4967]: I1008 00:07:53.708070 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:53 crc kubenswrapper[4967]: I1008 00:07:53.708082 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:53 crc kubenswrapper[4967]: I1008 00:07:53.708099 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:53 crc kubenswrapper[4967]: I1008 00:07:53.708111 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:53Z","lastTransitionTime":"2025-10-08T00:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:53 crc kubenswrapper[4967]: I1008 00:07:53.713628 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fa4b747-ab6e-4a11-9903-ef2dcb9668c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbadcdade274b6d9be940f22bc949e3c55cadf95dcad43c8cb7a1085c0e6708c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d0a0b2ef4a98ee828688081a74f4b7ecc31c8fae415d7257a560ecd43960ce3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96b53b1b283025b3d078a3f505d469d1872ad4bb8ab8e612c36c70b0b00a920\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef86a2422616e1e0d5c3afe844818c2bdf2150f941d1aefc506995f58587b2bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1400e110997c64fa33d6b8672c7e91cab5f782952ddf8e1dea2a882689026e7f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 00:06:56.760088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 00:06:56.761210 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-369805205/tls.crt::/tmp/serving-cert-369805205/tls.key\\\\\\\"\\\\nI1008 00:07:02.321448 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 00:07:02.325932 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 00:07:02.325980 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 00:07:02.326025 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 00:07:02.326042 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 00:07:02.336203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 00:07:02.336240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 00:07:02.336245 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 00:07:02.336250 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 00:07:02.336253 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 00:07:02.336256 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 00:07:02.336258 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 00:07:02.336430 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 00:07:02.340774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:46Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca28d8db19036e47a7ef3dadb70941f6c6cf5735e7703977d036185c46fce40\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5b222ad84c31d8746dfe03124c7287b401deb3ca2c62ce2b4fbd40e638cee0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5b222ad84c31d8746dfe03124c7287b401deb3ca2c62ce2b4fbd40e638cee0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:53Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:53 crc kubenswrapper[4967]: I1008 00:07:53.727058 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc98bc5e-fbae-4c67-a5c3-9738b3e966cd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acfae51bf8bee1b379a3859b7fb98ad0df3bed5d49dc0abe82b917512955c009\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae14677731c53e87871129d2140a34b18f0a48555309b652ebd490996fb1a723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae14677731c53e87871129d2140a34b18f0a48555309b652ebd490996fb1a723\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:53Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:53 crc kubenswrapper[4967]: I1008 00:07:53.741323 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:53Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:53 crc kubenswrapper[4967]: I1008 00:07:53.759460 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f1af2ba-8907-4d14-afa1-ffde207c80f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4535cffb8c9d64e662e282ecb44f6817b4604866f0bb75015f3b888d5554f821\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b03b83ddff429dbfa8fe052901aaa86a39ff0dd75477ffa8a6b6f4198f5819ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c9fa0559eb35ba67e11e36069ac8d0b2fec76399788eef668a08ad4353df3ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3312263e12bd42fbed0476ab5ace66b5497f62c9477865e130d67e3922e05cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e69dc603055aa221d44d11c049ddcc50cec568979911a310ed773a90341a499\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115f7a33f3dedae8f59d2f45f465f95f48b99c824e895ef7a0be9fe7a1b61d23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1cfd26b02d0e17b111b8ad84dab6c2adf692de66ff77158ee7cbff941c66ecd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1cfd26b02d0e17b111b8ad84dab6c2adf692de66ff77158ee7cbff941c66ecd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T00:07:29Z\\\",\\\"message\\\":\\\"iled objects of type *v1.Pod\\\\nI1008 00:07:28.973629 6638 obj_retry.go:409] Going to retry *v1.Pod resource setup for 1 objects: [openshift-multus/network-metrics-daemon-2h7z8]\\\\nI1008 00:07:28.973646 6638 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1008 00:07:28.973685 6638 obj_retry.go:285] Attempting retry of *v1.Pod openshift-multus/network-metrics-daemon-2h7z8 before timer (time: 2025-10-08 00:07:30.377116097 +0000 UTC m=+1.948919645): skip\\\\nI1008 00:07:28.973706 6638 obj_retry.go:420] Function iterateRetryResources for *v1.Pod ended (in 79.172µs)\\\\nI1008 00:07:28.973904 6638 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1008 00:07:28.973958 6638 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1008 00:07:28.974084 6638 factory.go:656] Stopping watch factory\\\\nI1008 00:07:28.974107 6638 ovnkube.go:599] Stopped ovnkube\\\\nI1008 00:07:28.974137 6638 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1008 00:07:28.974149 6638 handler.go:208] Removed *v1.Node event handler 2\\\\nI1008 00:07:28.974163 6638 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1008 00:07:28.974267 6638 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-bs8w9_openshift-ovn-kubernetes(4f1af2ba-8907-4d14-afa1-ffde207c80f4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27fb2b1225e386dcd7fecaaa0aa54ba1fdfafabfde6c3824ae07315313067542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bs8w9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:53Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:53 crc kubenswrapper[4967]: I1008 00:07:53.770321 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-v5n7t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abe7ae2d-7e63-4249-96e1-4000e9628ccc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52575e853ff6727dde7c89e4add4362be126127db27191c2a7bf2871efa719d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpsjd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:06Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-v5n7t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:53Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:53 crc kubenswrapper[4967]: I1008 00:07:53.781291 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94e63bc5641ce80d507f76a19b203f78c42368a7a4105f3fd09e19af21a1b735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:53Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:53 crc kubenswrapper[4967]: I1008 00:07:53.792244 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e98666fd3152bab57d3723e7da72aae2703b79f7ca73eb3c3e5f0e062a82592\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cf1cbe5548146fe359aafa351864125ec6b8bdd7b796aa4d311f6f3a16ad212\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:53Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:53 crc kubenswrapper[4967]: I1008 00:07:53.802422 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9tdbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41264a12-9486-4fee-a08c-607b590e5a70\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://806e85df796ae300d279f73a0cc4433a5bb1f7620ee8ce33c8dc4674c4ce0596\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5blv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9tdbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:53Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:53 crc kubenswrapper[4967]: I1008 00:07:53.810500 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:53 crc kubenswrapper[4967]: I1008 00:07:53.810783 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:53 crc kubenswrapper[4967]: I1008 00:07:53.810948 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:53 crc kubenswrapper[4967]: I1008 00:07:53.811102 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:53 crc kubenswrapper[4967]: I1008 00:07:53.811268 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:53Z","lastTransitionTime":"2025-10-08T00:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:53 crc kubenswrapper[4967]: I1008 00:07:53.812377 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2h7z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59f68771-e25d-4367-87ee-68a31d767882\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7r2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7r2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:18Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2h7z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:53Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:53 crc kubenswrapper[4967]: I1008 00:07:53.824762 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6ea2440-b2ae-481e-9442-0b3dda9688ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37fd5dcdfb864320d4d17729ab48c6da505fd867f384ad70a6dcf5bed78a3978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c7893dfaa0370f159d41c30cd869bb18550a819e8c3d20dd6f57dcbcc697a57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c430979d298477b4394c3567bb3800ab7dea874cee8fa41a6a50a19a7e3a0133\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://426e86ab409776ff0fa9505dc0d8ec36f27a5bf21d3aedafb19d3935bdba7529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://426e86ab409776ff0fa9505dc0d8ec36f27a5bf21d3aedafb19d3935bdba7529\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:53Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:53 crc kubenswrapper[4967]: I1008 00:07:53.836135 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:53Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:53 crc kubenswrapper[4967]: I1008 00:07:53.846526 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-npz2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c71501b1-aa9e-420e-b1bd-7e15365e7adc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9010a5d6165a64f986f9101dbbbcaf8a7d3fa5af85c14be4bf3aa8ed8536ad4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82dfe6c420d804b2d7daef4b5cfe3f87b77b369287c4d65553fa93f5f24f589d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T00:07:51Z\\\",\\\"message\\\":\\\"2025-10-08T00:07:06+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_b0e5397b-daec-448a-8e92-e00fdfae8aea\\\\n2025-10-08T00:07:06+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_b0e5397b-daec-448a-8e92-e00fdfae8aea to /host/opt/cni/bin/\\\\n2025-10-08T00:07:06Z [verbose] multus-daemon started\\\\n2025-10-08T00:07:06Z [verbose] Readiness Indicator file check\\\\n2025-10-08T00:07:51Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gblk2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-npz2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:53Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:53 crc kubenswrapper[4967]: I1008 00:07:53.913505 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:53 crc kubenswrapper[4967]: I1008 00:07:53.913582 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:53 crc kubenswrapper[4967]: I1008 00:07:53.913600 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:53 crc kubenswrapper[4967]: I1008 00:07:53.913625 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:53 crc kubenswrapper[4967]: I1008 00:07:53.913642 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:53Z","lastTransitionTime":"2025-10-08T00:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:54 crc kubenswrapper[4967]: I1008 00:07:54.016500 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:54 crc kubenswrapper[4967]: I1008 00:07:54.016757 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:54 crc kubenswrapper[4967]: I1008 00:07:54.016893 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:54 crc kubenswrapper[4967]: I1008 00:07:54.016957 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:54 crc kubenswrapper[4967]: I1008 00:07:54.017020 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:54Z","lastTransitionTime":"2025-10-08T00:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:54 crc kubenswrapper[4967]: I1008 00:07:54.117252 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2h7z8" Oct 08 00:07:54 crc kubenswrapper[4967]: I1008 00:07:54.117296 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 00:07:54 crc kubenswrapper[4967]: E1008 00:07:54.117598 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2h7z8" podUID="59f68771-e25d-4367-87ee-68a31d767882" Oct 08 00:07:54 crc kubenswrapper[4967]: E1008 00:07:54.117759 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 00:07:54 crc kubenswrapper[4967]: I1008 00:07:54.118926 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:54 crc kubenswrapper[4967]: I1008 00:07:54.118964 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:54 crc kubenswrapper[4967]: I1008 00:07:54.119041 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:54 crc kubenswrapper[4967]: I1008 00:07:54.119062 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:54 crc kubenswrapper[4967]: I1008 00:07:54.119074 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:54Z","lastTransitionTime":"2025-10-08T00:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:54 crc kubenswrapper[4967]: I1008 00:07:54.119080 4967 scope.go:117] "RemoveContainer" containerID="d1cfd26b02d0e17b111b8ad84dab6c2adf692de66ff77158ee7cbff941c66ecd" Oct 08 00:07:54 crc kubenswrapper[4967]: I1008 00:07:54.221724 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:54 crc kubenswrapper[4967]: I1008 00:07:54.221770 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:54 crc kubenswrapper[4967]: I1008 00:07:54.221783 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:54 crc kubenswrapper[4967]: I1008 00:07:54.221801 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:54 crc kubenswrapper[4967]: I1008 00:07:54.221814 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:54Z","lastTransitionTime":"2025-10-08T00:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:54 crc kubenswrapper[4967]: I1008 00:07:54.323951 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:54 crc kubenswrapper[4967]: I1008 00:07:54.323985 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:54 crc kubenswrapper[4967]: I1008 00:07:54.323997 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:54 crc kubenswrapper[4967]: I1008 00:07:54.324029 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:54 crc kubenswrapper[4967]: I1008 00:07:54.324045 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:54Z","lastTransitionTime":"2025-10-08T00:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:54 crc kubenswrapper[4967]: I1008 00:07:54.426371 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:54 crc kubenswrapper[4967]: I1008 00:07:54.426424 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:54 crc kubenswrapper[4967]: I1008 00:07:54.426446 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:54 crc kubenswrapper[4967]: I1008 00:07:54.426474 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:54 crc kubenswrapper[4967]: I1008 00:07:54.426495 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:54Z","lastTransitionTime":"2025-10-08T00:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:54 crc kubenswrapper[4967]: I1008 00:07:54.528570 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:54 crc kubenswrapper[4967]: I1008 00:07:54.528614 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:54 crc kubenswrapper[4967]: I1008 00:07:54.528624 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:54 crc kubenswrapper[4967]: I1008 00:07:54.528640 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:54 crc kubenswrapper[4967]: I1008 00:07:54.528649 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:54Z","lastTransitionTime":"2025-10-08T00:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:54 crc kubenswrapper[4967]: I1008 00:07:54.626301 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bs8w9_4f1af2ba-8907-4d14-afa1-ffde207c80f4/ovnkube-controller/2.log" Oct 08 00:07:54 crc kubenswrapper[4967]: I1008 00:07:54.628821 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" event={"ID":"4f1af2ba-8907-4d14-afa1-ffde207c80f4","Type":"ContainerStarted","Data":"4f7d46ee97fd3c4286f45879fededea788052cf83ebeebcb539be452f2ae102d"} Oct 08 00:07:54 crc kubenswrapper[4967]: I1008 00:07:54.629283 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" Oct 08 00:07:54 crc kubenswrapper[4967]: I1008 00:07:54.630110 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:54 crc kubenswrapper[4967]: I1008 00:07:54.630149 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:54 crc kubenswrapper[4967]: I1008 00:07:54.630159 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:54 crc kubenswrapper[4967]: I1008 00:07:54.630176 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:54 crc kubenswrapper[4967]: I1008 00:07:54.630186 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:54Z","lastTransitionTime":"2025-10-08T00:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:54 crc kubenswrapper[4967]: I1008 00:07:54.644380 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6ea2440-b2ae-481e-9442-0b3dda9688ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37fd5dcdfb864320d4d17729ab48c6da505fd867f384ad70a6dcf5bed78a3978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c7893dfaa0370f159d41c30cd869bb18550a819e8c3d20dd6f57dcbcc697a57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c430979d298477b4394c3567bb3800ab7dea874cee8fa41a6a50a19a7e3a0133\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://426e86ab409776ff0fa9505dc0d8ec36f27a5bf21d3aedafb19d3935bdba7529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://426e86ab409776ff0fa9505dc0d8ec36f27a5bf21d3aedafb19d3935bdba7529\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:54Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:54 crc kubenswrapper[4967]: I1008 00:07:54.667832 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:54Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:54 crc kubenswrapper[4967]: I1008 00:07:54.682937 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-npz2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c71501b1-aa9e-420e-b1bd-7e15365e7adc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9010a5d6165a64f986f9101dbbbcaf8a7d3fa5af85c14be4bf3aa8ed8536ad4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82dfe6c420d804b2d7daef4b5cfe3f87b77b369287c4d65553fa93f5f24f589d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T00:07:51Z\\\",\\\"message\\\":\\\"2025-10-08T00:07:06+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_b0e5397b-daec-448a-8e92-e00fdfae8aea\\\\n2025-10-08T00:07:06+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_b0e5397b-daec-448a-8e92-e00fdfae8aea to /host/opt/cni/bin/\\\\n2025-10-08T00:07:06Z [verbose] multus-daemon started\\\\n2025-10-08T00:07:06Z [verbose] Readiness Indicator file check\\\\n2025-10-08T00:07:51Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gblk2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-npz2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:54Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:54 crc kubenswrapper[4967]: I1008 00:07:54.706991 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2h7z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59f68771-e25d-4367-87ee-68a31d767882\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7r2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7r2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:18Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2h7z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:54Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:54 crc kubenswrapper[4967]: I1008 00:07:54.728505 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58c4704-a88a-43b4-b944-4de626f13f7e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22f5dabb39da0a7bd01db4dae1437b9a231a1fcc0dfe749a4a5a637a72cf33c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01a2b3dc785520c79eae998482e5a1347725c76584357073238fe60a619edb77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6aa6f5e81feceab3d1c7587152a70546b365e629a69267e4d3084324a1de764a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63c2b7c4b2791bd1f16888c58b878dbff1f87884e68c9949de914d0d3e1d39fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:54Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:54 crc kubenswrapper[4967]: I1008 00:07:54.731640 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:54 crc kubenswrapper[4967]: I1008 00:07:54.731705 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:54 crc kubenswrapper[4967]: I1008 00:07:54.731719 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:54 crc kubenswrapper[4967]: I1008 00:07:54.731735 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:54 crc kubenswrapper[4967]: I1008 00:07:54.731746 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:54Z","lastTransitionTime":"2025-10-08T00:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:54 crc kubenswrapper[4967]: I1008 00:07:54.738093 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:54Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:54 crc kubenswrapper[4967]: I1008 00:07:54.750749 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9mrr7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9299bc4f-f8d3-4baf-9885-39573982fd64\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://994beafece9dead0d85ae74dc453a142460705fd66b52ba22f2fa799cf27da08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb49c8476b482416588c4d403016940730064a619c0641eb029c947a4bab1215\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb49c8476b482416588c4d403016940730064a619c0641eb029c947a4bab1215\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b7b63962dcbf477f1fa4d71826a4aa3358920a2c1b2f5e48716411b5bdca049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b7b63962dcbf477f1fa4d71826a4aa3358920a2c1b2f5e48716411b5bdca049\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1e770dc737173fd4f3220ed6c52b422ccbd829722386a37029e1c3e10d80e1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1e770dc737173fd4f3220ed6c52b422ccbd829722386a37029e1c3e10d80e1a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e357d970878c996f219a79f1876b363176c9847b9575f33090f62bc299529a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e357d970878c996f219a79f1876b363176c9847b9575f33090f62bc299529a95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f098966bfad49d06278eaa72be042596d488115e597737d9a0a0cfc78f99f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49f098966bfad49d06278eaa72be042596d488115e597737d9a0a0cfc78f99f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1d53876b7b863c612d597fd716c3260e8857b6542802073bb730f6756071b46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c1d53876b7b863c612d597fd716c3260e8857b6542802073bb730f6756071b46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9mrr7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:54Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:54 crc kubenswrapper[4967]: I1008 00:07:54.760385 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kpmlg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f975d47d-42a6-422e-80e2-d1ee282817fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a317ae6640eb38b3aee0f9738cca890e8ab3d92075b620cd312f40a0628d46c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qznqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ef61854aea40924c5372b0553cd0b7e8faa39a1d42fec5b65cf867c368f72c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qznqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-kpmlg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:54Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:54 crc kubenswrapper[4967]: I1008 00:07:54.772851 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fa4b747-ab6e-4a11-9903-ef2dcb9668c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbadcdade274b6d9be940f22bc949e3c55cadf95dcad43c8cb7a1085c0e6708c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d0a0b2ef4a98ee828688081a74f4b7ecc31c8fae415d7257a560ecd43960ce3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96b53b1b283025b3d078a3f505d469d1872ad4bb8ab8e612c36c70b0b00a920\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef86a2422616e1e0d5c3afe844818c2bdf2150f941d1aefc506995f58587b2bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1400e110997c64fa33d6b8672c7e91cab5f782952ddf8e1dea2a882689026e7f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 00:06:56.760088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 00:06:56.761210 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-369805205/tls.crt::/tmp/serving-cert-369805205/tls.key\\\\\\\"\\\\nI1008 00:07:02.321448 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 00:07:02.325932 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 00:07:02.325980 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 00:07:02.326025 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 00:07:02.326042 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 00:07:02.336203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 00:07:02.336240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 00:07:02.336245 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 00:07:02.336250 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 00:07:02.336253 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 00:07:02.336256 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 00:07:02.336258 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 00:07:02.336430 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 00:07:02.340774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:46Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca28d8db19036e47a7ef3dadb70941f6c6cf5735e7703977d036185c46fce40\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5b222ad84c31d8746dfe03124c7287b401deb3ca2c62ce2b4fbd40e638cee0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5b222ad84c31d8746dfe03124c7287b401deb3ca2c62ce2b4fbd40e638cee0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:54Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:54 crc kubenswrapper[4967]: I1008 00:07:54.784960 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc98bc5e-fbae-4c67-a5c3-9738b3e966cd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acfae51bf8bee1b379a3859b7fb98ad0df3bed5d49dc0abe82b917512955c009\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae14677731c53e87871129d2140a34b18f0a48555309b652ebd490996fb1a723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae14677731c53e87871129d2140a34b18f0a48555309b652ebd490996fb1a723\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:54Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:54 crc kubenswrapper[4967]: I1008 00:07:54.832396 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:54Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:54 crc kubenswrapper[4967]: I1008 00:07:54.833387 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:54 crc kubenswrapper[4967]: I1008 00:07:54.833411 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:54 crc kubenswrapper[4967]: I1008 00:07:54.833420 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:54 crc kubenswrapper[4967]: I1008 00:07:54.833433 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:54 crc kubenswrapper[4967]: I1008 00:07:54.833442 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:54Z","lastTransitionTime":"2025-10-08T00:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:54 crc kubenswrapper[4967]: I1008 00:07:54.845686 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbfcdc882e2be1cb9cf666c9e7775671078f7dbbfe8421dd15db737b89920a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:54Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:54 crc kubenswrapper[4967]: I1008 00:07:54.856403 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9461162a-c27d-465e-9ca7-2d3936877cd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9778d78ddc7325abd4d77bc5e71b0b114f50be5dda4a438c7cb6b05640c8a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z9tp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28760af04ef321d0f91b2bce3ad074d44b63dabf676cc37f3be6aa8726341a2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z9tp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xvdtp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:54Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:54 crc kubenswrapper[4967]: I1008 00:07:54.866859 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94e63bc5641ce80d507f76a19b203f78c42368a7a4105f3fd09e19af21a1b735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:54Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:54 crc kubenswrapper[4967]: I1008 00:07:54.886145 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e98666fd3152bab57d3723e7da72aae2703b79f7ca73eb3c3e5f0e062a82592\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cf1cbe5548146fe359aafa351864125ec6b8bdd7b796aa4d311f6f3a16ad212\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:54Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:54 crc kubenswrapper[4967]: I1008 00:07:54.894164 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9tdbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41264a12-9486-4fee-a08c-607b590e5a70\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://806e85df796ae300d279f73a0cc4433a5bb1f7620ee8ce33c8dc4674c4ce0596\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5blv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9tdbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:54Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:54 crc kubenswrapper[4967]: I1008 00:07:54.911681 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f1af2ba-8907-4d14-afa1-ffde207c80f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4535cffb8c9d64e662e282ecb44f6817b4604866f0bb75015f3b888d5554f821\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b03b83ddff429dbfa8fe052901aaa86a39ff0dd75477ffa8a6b6f4198f5819ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c9fa0559eb35ba67e11e36069ac8d0b2fec76399788eef668a08ad4353df3ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3312263e12bd42fbed0476ab5ace66b5497f62c9477865e130d67e3922e05cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e69dc603055aa221d44d11c049ddcc50cec568979911a310ed773a90341a499\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115f7a33f3dedae8f59d2f45f465f95f48b99c824e895ef7a0be9fe7a1b61d23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f7d46ee97fd3c4286f45879fededea788052cf83ebeebcb539be452f2ae102d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1cfd26b02d0e17b111b8ad84dab6c2adf692de66ff77158ee7cbff941c66ecd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T00:07:29Z\\\",\\\"message\\\":\\\"iled objects of type *v1.Pod\\\\nI1008 00:07:28.973629 6638 obj_retry.go:409] Going to retry *v1.Pod resource setup for 1 objects: [openshift-multus/network-metrics-daemon-2h7z8]\\\\nI1008 00:07:28.973646 6638 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1008 00:07:28.973685 6638 obj_retry.go:285] Attempting retry of *v1.Pod openshift-multus/network-metrics-daemon-2h7z8 before timer (time: 2025-10-08 00:07:30.377116097 +0000 UTC m=+1.948919645): skip\\\\nI1008 00:07:28.973706 6638 obj_retry.go:420] Function iterateRetryResources for *v1.Pod ended (in 79.172µs)\\\\nI1008 00:07:28.973904 6638 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1008 00:07:28.973958 6638 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1008 00:07:28.974084 6638 factory.go:656] Stopping watch factory\\\\nI1008 00:07:28.974107 6638 ovnkube.go:599] Stopped ovnkube\\\\nI1008 00:07:28.974137 6638 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1008 00:07:28.974149 6638 handler.go:208] Removed *v1.Node event handler 2\\\\nI1008 00:07:28.974163 6638 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1008 00:07:28.974267 6638 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27fb2b1225e386dcd7fecaaa0aa54ba1fdfafabfde6c3824ae07315313067542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bs8w9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:54Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:54 crc kubenswrapper[4967]: I1008 00:07:54.922509 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-v5n7t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abe7ae2d-7e63-4249-96e1-4000e9628ccc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52575e853ff6727dde7c89e4add4362be126127db27191c2a7bf2871efa719d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpsjd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:06Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-v5n7t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:54Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:54 crc kubenswrapper[4967]: I1008 00:07:54.936401 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:54 crc kubenswrapper[4967]: I1008 00:07:54.936450 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:54 crc kubenswrapper[4967]: I1008 00:07:54.936462 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:54 crc kubenswrapper[4967]: I1008 00:07:54.936479 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:54 crc kubenswrapper[4967]: I1008 00:07:54.936492 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:54Z","lastTransitionTime":"2025-10-08T00:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:55 crc kubenswrapper[4967]: I1008 00:07:55.039110 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:55 crc kubenswrapper[4967]: I1008 00:07:55.039149 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:55 crc kubenswrapper[4967]: I1008 00:07:55.039157 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:55 crc kubenswrapper[4967]: I1008 00:07:55.039174 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:55 crc kubenswrapper[4967]: I1008 00:07:55.039185 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:55Z","lastTransitionTime":"2025-10-08T00:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:55 crc kubenswrapper[4967]: I1008 00:07:55.117193 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 00:07:55 crc kubenswrapper[4967]: I1008 00:07:55.117208 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 00:07:55 crc kubenswrapper[4967]: E1008 00:07:55.117340 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 00:07:55 crc kubenswrapper[4967]: E1008 00:07:55.117426 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 00:07:55 crc kubenswrapper[4967]: I1008 00:07:55.142047 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:55 crc kubenswrapper[4967]: I1008 00:07:55.142088 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:55 crc kubenswrapper[4967]: I1008 00:07:55.142099 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:55 crc kubenswrapper[4967]: I1008 00:07:55.142114 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:55 crc kubenswrapper[4967]: I1008 00:07:55.142123 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:55Z","lastTransitionTime":"2025-10-08T00:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:55 crc kubenswrapper[4967]: I1008 00:07:55.245877 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:55 crc kubenswrapper[4967]: I1008 00:07:55.245929 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:55 crc kubenswrapper[4967]: I1008 00:07:55.245939 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:55 crc kubenswrapper[4967]: I1008 00:07:55.245954 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:55 crc kubenswrapper[4967]: I1008 00:07:55.245963 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:55Z","lastTransitionTime":"2025-10-08T00:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:55 crc kubenswrapper[4967]: I1008 00:07:55.347969 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:55 crc kubenswrapper[4967]: I1008 00:07:55.348018 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:55 crc kubenswrapper[4967]: I1008 00:07:55.348032 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:55 crc kubenswrapper[4967]: I1008 00:07:55.348046 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:55 crc kubenswrapper[4967]: I1008 00:07:55.348061 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:55Z","lastTransitionTime":"2025-10-08T00:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:55 crc kubenswrapper[4967]: I1008 00:07:55.451051 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:55 crc kubenswrapper[4967]: I1008 00:07:55.451099 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:55 crc kubenswrapper[4967]: I1008 00:07:55.451111 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:55 crc kubenswrapper[4967]: I1008 00:07:55.451128 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:55 crc kubenswrapper[4967]: I1008 00:07:55.451139 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:55Z","lastTransitionTime":"2025-10-08T00:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:55 crc kubenswrapper[4967]: I1008 00:07:55.553414 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:55 crc kubenswrapper[4967]: I1008 00:07:55.553453 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:55 crc kubenswrapper[4967]: I1008 00:07:55.553464 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:55 crc kubenswrapper[4967]: I1008 00:07:55.553478 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:55 crc kubenswrapper[4967]: I1008 00:07:55.553487 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:55Z","lastTransitionTime":"2025-10-08T00:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:55 crc kubenswrapper[4967]: I1008 00:07:55.636473 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bs8w9_4f1af2ba-8907-4d14-afa1-ffde207c80f4/ovnkube-controller/3.log" Oct 08 00:07:55 crc kubenswrapper[4967]: I1008 00:07:55.637466 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bs8w9_4f1af2ba-8907-4d14-afa1-ffde207c80f4/ovnkube-controller/2.log" Oct 08 00:07:55 crc kubenswrapper[4967]: I1008 00:07:55.640280 4967 generic.go:334] "Generic (PLEG): container finished" podID="4f1af2ba-8907-4d14-afa1-ffde207c80f4" containerID="4f7d46ee97fd3c4286f45879fededea788052cf83ebeebcb539be452f2ae102d" exitCode=1 Oct 08 00:07:55 crc kubenswrapper[4967]: I1008 00:07:55.640312 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" event={"ID":"4f1af2ba-8907-4d14-afa1-ffde207c80f4","Type":"ContainerDied","Data":"4f7d46ee97fd3c4286f45879fededea788052cf83ebeebcb539be452f2ae102d"} Oct 08 00:07:55 crc kubenswrapper[4967]: I1008 00:07:55.640377 4967 scope.go:117] "RemoveContainer" containerID="d1cfd26b02d0e17b111b8ad84dab6c2adf692de66ff77158ee7cbff941c66ecd" Oct 08 00:07:55 crc kubenswrapper[4967]: I1008 00:07:55.641973 4967 scope.go:117] "RemoveContainer" containerID="4f7d46ee97fd3c4286f45879fededea788052cf83ebeebcb539be452f2ae102d" Oct 08 00:07:55 crc kubenswrapper[4967]: E1008 00:07:55.642129 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-bs8w9_openshift-ovn-kubernetes(4f1af2ba-8907-4d14-afa1-ffde207c80f4)\"" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" podUID="4f1af2ba-8907-4d14-afa1-ffde207c80f4" Oct 08 00:07:55 crc kubenswrapper[4967]: I1008 00:07:55.654758 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6ea2440-b2ae-481e-9442-0b3dda9688ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37fd5dcdfb864320d4d17729ab48c6da505fd867f384ad70a6dcf5bed78a3978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c7893dfaa0370f159d41c30cd869bb18550a819e8c3d20dd6f57dcbcc697a57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c430979d298477b4394c3567bb3800ab7dea874cee8fa41a6a50a19a7e3a0133\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://426e86ab409776ff0fa9505dc0d8ec36f27a5bf21d3aedafb19d3935bdba7529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://426e86ab409776ff0fa9505dc0d8ec36f27a5bf21d3aedafb19d3935bdba7529\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:55Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:55 crc kubenswrapper[4967]: I1008 00:07:55.655716 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:55 crc kubenswrapper[4967]: I1008 00:07:55.655753 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:55 crc kubenswrapper[4967]: I1008 00:07:55.655766 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:55 crc kubenswrapper[4967]: I1008 00:07:55.655782 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:55 crc kubenswrapper[4967]: I1008 00:07:55.655793 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:55Z","lastTransitionTime":"2025-10-08T00:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:55 crc kubenswrapper[4967]: I1008 00:07:55.667469 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:55Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:55 crc kubenswrapper[4967]: I1008 00:07:55.680885 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-npz2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c71501b1-aa9e-420e-b1bd-7e15365e7adc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9010a5d6165a64f986f9101dbbbcaf8a7d3fa5af85c14be4bf3aa8ed8536ad4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82dfe6c420d804b2d7daef4b5cfe3f87b77b369287c4d65553fa93f5f24f589d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T00:07:51Z\\\",\\\"message\\\":\\\"2025-10-08T00:07:06+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_b0e5397b-daec-448a-8e92-e00fdfae8aea\\\\n2025-10-08T00:07:06+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_b0e5397b-daec-448a-8e92-e00fdfae8aea to /host/opt/cni/bin/\\\\n2025-10-08T00:07:06Z [verbose] multus-daemon started\\\\n2025-10-08T00:07:06Z [verbose] Readiness Indicator file check\\\\n2025-10-08T00:07:51Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gblk2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-npz2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:55Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:55 crc kubenswrapper[4967]: I1008 00:07:55.690215 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2h7z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59f68771-e25d-4367-87ee-68a31d767882\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7r2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7r2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:18Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2h7z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:55Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:55 crc kubenswrapper[4967]: I1008 00:07:55.702469 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kpmlg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f975d47d-42a6-422e-80e2-d1ee282817fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a317ae6640eb38b3aee0f9738cca890e8ab3d92075b620cd312f40a0628d46c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qznqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ef61854aea40924c5372b0553cd0b7e8faa39a1d42fec5b65cf867c368f72c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qznqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-kpmlg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:55Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:55 crc kubenswrapper[4967]: I1008 00:07:55.714495 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58c4704-a88a-43b4-b944-4de626f13f7e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22f5dabb39da0a7bd01db4dae1437b9a231a1fcc0dfe749a4a5a637a72cf33c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01a2b3dc785520c79eae998482e5a1347725c76584357073238fe60a619edb77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6aa6f5e81feceab3d1c7587152a70546b365e629a69267e4d3084324a1de764a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63c2b7c4b2791bd1f16888c58b878dbff1f87884e68c9949de914d0d3e1d39fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:55Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:55 crc kubenswrapper[4967]: I1008 00:07:55.726659 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:55Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:55 crc kubenswrapper[4967]: I1008 00:07:55.741204 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9mrr7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9299bc4f-f8d3-4baf-9885-39573982fd64\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://994beafece9dead0d85ae74dc453a142460705fd66b52ba22f2fa799cf27da08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb49c8476b482416588c4d403016940730064a619c0641eb029c947a4bab1215\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb49c8476b482416588c4d403016940730064a619c0641eb029c947a4bab1215\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b7b63962dcbf477f1fa4d71826a4aa3358920a2c1b2f5e48716411b5bdca049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b7b63962dcbf477f1fa4d71826a4aa3358920a2c1b2f5e48716411b5bdca049\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1e770dc737173fd4f3220ed6c52b422ccbd829722386a37029e1c3e10d80e1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1e770dc737173fd4f3220ed6c52b422ccbd829722386a37029e1c3e10d80e1a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e357d970878c996f219a79f1876b363176c9847b9575f33090f62bc299529a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e357d970878c996f219a79f1876b363176c9847b9575f33090f62bc299529a95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f098966bfad49d06278eaa72be042596d488115e597737d9a0a0cfc78f99f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49f098966bfad49d06278eaa72be042596d488115e597737d9a0a0cfc78f99f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1d53876b7b863c612d597fd716c3260e8857b6542802073bb730f6756071b46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c1d53876b7b863c612d597fd716c3260e8857b6542802073bb730f6756071b46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9mrr7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:55Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:55 crc kubenswrapper[4967]: I1008 00:07:55.755503 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9461162a-c27d-465e-9ca7-2d3936877cd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9778d78ddc7325abd4d77bc5e71b0b114f50be5dda4a438c7cb6b05640c8a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z9tp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28760af04ef321d0f91b2bce3ad074d44b63dabf676cc37f3be6aa8726341a2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z9tp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xvdtp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:55Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:55 crc kubenswrapper[4967]: I1008 00:07:55.758427 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:55 crc kubenswrapper[4967]: I1008 00:07:55.758486 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:55 crc kubenswrapper[4967]: I1008 00:07:55.758501 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:55 crc kubenswrapper[4967]: I1008 00:07:55.758522 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:55 crc kubenswrapper[4967]: I1008 00:07:55.758535 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:55Z","lastTransitionTime":"2025-10-08T00:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:55 crc kubenswrapper[4967]: I1008 00:07:55.771422 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fa4b747-ab6e-4a11-9903-ef2dcb9668c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbadcdade274b6d9be940f22bc949e3c55cadf95dcad43c8cb7a1085c0e6708c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d0a0b2ef4a98ee828688081a74f4b7ecc31c8fae415d7257a560ecd43960ce3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96b53b1b283025b3d078a3f505d469d1872ad4bb8ab8e612c36c70b0b00a920\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef86a2422616e1e0d5c3afe844818c2bdf2150f941d1aefc506995f58587b2bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1400e110997c64fa33d6b8672c7e91cab5f782952ddf8e1dea2a882689026e7f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 00:06:56.760088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 00:06:56.761210 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-369805205/tls.crt::/tmp/serving-cert-369805205/tls.key\\\\\\\"\\\\nI1008 00:07:02.321448 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 00:07:02.325932 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 00:07:02.325980 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 00:07:02.326025 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 00:07:02.326042 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 00:07:02.336203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 00:07:02.336240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 00:07:02.336245 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 00:07:02.336250 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 00:07:02.336253 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 00:07:02.336256 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 00:07:02.336258 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 00:07:02.336430 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 00:07:02.340774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:46Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca28d8db19036e47a7ef3dadb70941f6c6cf5735e7703977d036185c46fce40\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5b222ad84c31d8746dfe03124c7287b401deb3ca2c62ce2b4fbd40e638cee0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5b222ad84c31d8746dfe03124c7287b401deb3ca2c62ce2b4fbd40e638cee0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:55Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:55 crc kubenswrapper[4967]: I1008 00:07:55.781278 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc98bc5e-fbae-4c67-a5c3-9738b3e966cd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acfae51bf8bee1b379a3859b7fb98ad0df3bed5d49dc0abe82b917512955c009\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae14677731c53e87871129d2140a34b18f0a48555309b652ebd490996fb1a723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae14677731c53e87871129d2140a34b18f0a48555309b652ebd490996fb1a723\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:55Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:55 crc kubenswrapper[4967]: I1008 00:07:55.792876 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:55Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:55 crc kubenswrapper[4967]: I1008 00:07:55.807066 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbfcdc882e2be1cb9cf666c9e7775671078f7dbbfe8421dd15db737b89920a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:55Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:55 crc kubenswrapper[4967]: I1008 00:07:55.820126 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-v5n7t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abe7ae2d-7e63-4249-96e1-4000e9628ccc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52575e853ff6727dde7c89e4add4362be126127db27191c2a7bf2871efa719d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpsjd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:06Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-v5n7t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:55Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:55 crc kubenswrapper[4967]: I1008 00:07:55.833038 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94e63bc5641ce80d507f76a19b203f78c42368a7a4105f3fd09e19af21a1b735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:55Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:55 crc kubenswrapper[4967]: I1008 00:07:55.843971 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e98666fd3152bab57d3723e7da72aae2703b79f7ca73eb3c3e5f0e062a82592\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cf1cbe5548146fe359aafa351864125ec6b8bdd7b796aa4d311f6f3a16ad212\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:55Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:55 crc kubenswrapper[4967]: I1008 00:07:55.853297 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9tdbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41264a12-9486-4fee-a08c-607b590e5a70\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://806e85df796ae300d279f73a0cc4433a5bb1f7620ee8ce33c8dc4674c4ce0596\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5blv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9tdbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:55Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:55 crc kubenswrapper[4967]: I1008 00:07:55.860860 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:55 crc kubenswrapper[4967]: I1008 00:07:55.860892 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:55 crc kubenswrapper[4967]: I1008 00:07:55.860901 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:55 crc kubenswrapper[4967]: I1008 00:07:55.860913 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:55 crc kubenswrapper[4967]: I1008 00:07:55.860922 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:55Z","lastTransitionTime":"2025-10-08T00:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:55 crc kubenswrapper[4967]: I1008 00:07:55.871819 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f1af2ba-8907-4d14-afa1-ffde207c80f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4535cffb8c9d64e662e282ecb44f6817b4604866f0bb75015f3b888d5554f821\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b03b83ddff429dbfa8fe052901aaa86a39ff0dd75477ffa8a6b6f4198f5819ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c9fa0559eb35ba67e11e36069ac8d0b2fec76399788eef668a08ad4353df3ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3312263e12bd42fbed0476ab5ace66b5497f62c9477865e130d67e3922e05cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e69dc603055aa221d44d11c049ddcc50cec568979911a310ed773a90341a499\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115f7a33f3dedae8f59d2f45f465f95f48b99c824e895ef7a0be9fe7a1b61d23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f7d46ee97fd3c4286f45879fededea788052cf83ebeebcb539be452f2ae102d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1cfd26b02d0e17b111b8ad84dab6c2adf692de66ff77158ee7cbff941c66ecd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T00:07:29Z\\\",\\\"message\\\":\\\"iled objects of type *v1.Pod\\\\nI1008 00:07:28.973629 6638 obj_retry.go:409] Going to retry *v1.Pod resource setup for 1 objects: [openshift-multus/network-metrics-daemon-2h7z8]\\\\nI1008 00:07:28.973646 6638 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1008 00:07:28.973685 6638 obj_retry.go:285] Attempting retry of *v1.Pod openshift-multus/network-metrics-daemon-2h7z8 before timer (time: 2025-10-08 00:07:30.377116097 +0000 UTC m=+1.948919645): skip\\\\nI1008 00:07:28.973706 6638 obj_retry.go:420] Function iterateRetryResources for *v1.Pod ended (in 79.172µs)\\\\nI1008 00:07:28.973904 6638 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1008 00:07:28.973958 6638 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1008 00:07:28.974084 6638 factory.go:656] Stopping watch factory\\\\nI1008 00:07:28.974107 6638 ovnkube.go:599] Stopped ovnkube\\\\nI1008 00:07:28.974137 6638 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1008 00:07:28.974149 6638 handler.go:208] Removed *v1.Node event handler 2\\\\nI1008 00:07:28.974163 6638 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1008 00:07:28.974267 6638 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:28Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f7d46ee97fd3c4286f45879fededea788052cf83ebeebcb539be452f2ae102d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T00:07:55Z\\\",\\\"message\\\":\\\"ue, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.53\\\\\\\", Port:8383, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.53\\\\\\\", Port:8081, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1008 00:07:55.166176 6981 services_controller.go:452] Built service openshift-marketplace/marketplace-operator-metrics per-node LB for network=default: []services.LB{}\\\\nI1008 00:07:55.166184 6981 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:960d98b2-dc64-4e93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1008 00:07:55.165765 6981 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, h\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27fb2b1225e386dcd7fecaaa0aa54ba1fdfafabfde6c3824ae07315313067542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bs8w9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:55Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:55 crc kubenswrapper[4967]: I1008 00:07:55.962857 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:55 crc kubenswrapper[4967]: I1008 00:07:55.962885 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:55 crc kubenswrapper[4967]: I1008 00:07:55.962894 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:55 crc kubenswrapper[4967]: I1008 00:07:55.962906 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:55 crc kubenswrapper[4967]: I1008 00:07:55.962914 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:55Z","lastTransitionTime":"2025-10-08T00:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:56 crc kubenswrapper[4967]: I1008 00:07:56.064491 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:56 crc kubenswrapper[4967]: I1008 00:07:56.064532 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:56 crc kubenswrapper[4967]: I1008 00:07:56.064544 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:56 crc kubenswrapper[4967]: I1008 00:07:56.064559 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:56 crc kubenswrapper[4967]: I1008 00:07:56.064579 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:56Z","lastTransitionTime":"2025-10-08T00:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:56 crc kubenswrapper[4967]: I1008 00:07:56.117762 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2h7z8" Oct 08 00:07:56 crc kubenswrapper[4967]: I1008 00:07:56.117854 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 00:07:56 crc kubenswrapper[4967]: E1008 00:07:56.118026 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2h7z8" podUID="59f68771-e25d-4367-87ee-68a31d767882" Oct 08 00:07:56 crc kubenswrapper[4967]: E1008 00:07:56.118402 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 00:07:56 crc kubenswrapper[4967]: I1008 00:07:56.167070 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:56 crc kubenswrapper[4967]: I1008 00:07:56.167103 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:56 crc kubenswrapper[4967]: I1008 00:07:56.167186 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:56 crc kubenswrapper[4967]: I1008 00:07:56.167205 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:56 crc kubenswrapper[4967]: I1008 00:07:56.167214 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:56Z","lastTransitionTime":"2025-10-08T00:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:56 crc kubenswrapper[4967]: I1008 00:07:56.269946 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:56 crc kubenswrapper[4967]: I1008 00:07:56.269997 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:56 crc kubenswrapper[4967]: I1008 00:07:56.270018 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:56 crc kubenswrapper[4967]: I1008 00:07:56.270045 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:56 crc kubenswrapper[4967]: I1008 00:07:56.270067 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:56Z","lastTransitionTime":"2025-10-08T00:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:56 crc kubenswrapper[4967]: I1008 00:07:56.373416 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:56 crc kubenswrapper[4967]: I1008 00:07:56.373448 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:56 crc kubenswrapper[4967]: I1008 00:07:56.373459 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:56 crc kubenswrapper[4967]: I1008 00:07:56.373489 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:56 crc kubenswrapper[4967]: I1008 00:07:56.373500 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:56Z","lastTransitionTime":"2025-10-08T00:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:56 crc kubenswrapper[4967]: I1008 00:07:56.476700 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:56 crc kubenswrapper[4967]: I1008 00:07:56.476740 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:56 crc kubenswrapper[4967]: I1008 00:07:56.476753 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:56 crc kubenswrapper[4967]: I1008 00:07:56.476771 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:56 crc kubenswrapper[4967]: I1008 00:07:56.476783 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:56Z","lastTransitionTime":"2025-10-08T00:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:56 crc kubenswrapper[4967]: I1008 00:07:56.578544 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:56 crc kubenswrapper[4967]: I1008 00:07:56.578591 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:56 crc kubenswrapper[4967]: I1008 00:07:56.578608 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:56 crc kubenswrapper[4967]: I1008 00:07:56.578631 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:56 crc kubenswrapper[4967]: I1008 00:07:56.578648 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:56Z","lastTransitionTime":"2025-10-08T00:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:56 crc kubenswrapper[4967]: I1008 00:07:56.647111 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bs8w9_4f1af2ba-8907-4d14-afa1-ffde207c80f4/ovnkube-controller/3.log" Oct 08 00:07:56 crc kubenswrapper[4967]: I1008 00:07:56.652453 4967 scope.go:117] "RemoveContainer" containerID="4f7d46ee97fd3c4286f45879fededea788052cf83ebeebcb539be452f2ae102d" Oct 08 00:07:56 crc kubenswrapper[4967]: E1008 00:07:56.652814 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-bs8w9_openshift-ovn-kubernetes(4f1af2ba-8907-4d14-afa1-ffde207c80f4)\"" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" podUID="4f1af2ba-8907-4d14-afa1-ffde207c80f4" Oct 08 00:07:56 crc kubenswrapper[4967]: I1008 00:07:56.668426 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbfcdc882e2be1cb9cf666c9e7775671078f7dbbfe8421dd15db737b89920a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:56Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:56 crc kubenswrapper[4967]: I1008 00:07:56.682274 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:56 crc kubenswrapper[4967]: I1008 00:07:56.682427 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:56 crc kubenswrapper[4967]: I1008 00:07:56.682512 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:56 crc kubenswrapper[4967]: I1008 00:07:56.682600 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:56 crc kubenswrapper[4967]: I1008 00:07:56.682700 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:56Z","lastTransitionTime":"2025-10-08T00:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:56 crc kubenswrapper[4967]: I1008 00:07:56.683620 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9461162a-c27d-465e-9ca7-2d3936877cd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9778d78ddc7325abd4d77bc5e71b0b114f50be5dda4a438c7cb6b05640c8a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z9tp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28760af04ef321d0f91b2bce3ad074d44b63dabf676cc37f3be6aa8726341a2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z9tp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xvdtp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:56Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:56 crc kubenswrapper[4967]: I1008 00:07:56.700134 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fa4b747-ab6e-4a11-9903-ef2dcb9668c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbadcdade274b6d9be940f22bc949e3c55cadf95dcad43c8cb7a1085c0e6708c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d0a0b2ef4a98ee828688081a74f4b7ecc31c8fae415d7257a560ecd43960ce3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96b53b1b283025b3d078a3f505d469d1872ad4bb8ab8e612c36c70b0b00a920\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef86a2422616e1e0d5c3afe844818c2bdf2150f941d1aefc506995f58587b2bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1400e110997c64fa33d6b8672c7e91cab5f782952ddf8e1dea2a882689026e7f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 00:06:56.760088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 00:06:56.761210 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-369805205/tls.crt::/tmp/serving-cert-369805205/tls.key\\\\\\\"\\\\nI1008 00:07:02.321448 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 00:07:02.325932 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 00:07:02.325980 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 00:07:02.326025 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 00:07:02.326042 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 00:07:02.336203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 00:07:02.336240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 00:07:02.336245 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 00:07:02.336250 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 00:07:02.336253 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 00:07:02.336256 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 00:07:02.336258 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 00:07:02.336430 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 00:07:02.340774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:46Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca28d8db19036e47a7ef3dadb70941f6c6cf5735e7703977d036185c46fce40\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5b222ad84c31d8746dfe03124c7287b401deb3ca2c62ce2b4fbd40e638cee0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5b222ad84c31d8746dfe03124c7287b401deb3ca2c62ce2b4fbd40e638cee0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:56Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:56 crc kubenswrapper[4967]: I1008 00:07:56.714375 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc98bc5e-fbae-4c67-a5c3-9738b3e966cd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acfae51bf8bee1b379a3859b7fb98ad0df3bed5d49dc0abe82b917512955c009\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae14677731c53e87871129d2140a34b18f0a48555309b652ebd490996fb1a723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae14677731c53e87871129d2140a34b18f0a48555309b652ebd490996fb1a723\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:56Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:56 crc kubenswrapper[4967]: I1008 00:07:56.729406 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:56Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:56 crc kubenswrapper[4967]: I1008 00:07:56.750142 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f1af2ba-8907-4d14-afa1-ffde207c80f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4535cffb8c9d64e662e282ecb44f6817b4604866f0bb75015f3b888d5554f821\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b03b83ddff429dbfa8fe052901aaa86a39ff0dd75477ffa8a6b6f4198f5819ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c9fa0559eb35ba67e11e36069ac8d0b2fec76399788eef668a08ad4353df3ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3312263e12bd42fbed0476ab5ace66b5497f62c9477865e130d67e3922e05cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e69dc603055aa221d44d11c049ddcc50cec568979911a310ed773a90341a499\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115f7a33f3dedae8f59d2f45f465f95f48b99c824e895ef7a0be9fe7a1b61d23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f7d46ee97fd3c4286f45879fededea788052cf83ebeebcb539be452f2ae102d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f7d46ee97fd3c4286f45879fededea788052cf83ebeebcb539be452f2ae102d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T00:07:55Z\\\",\\\"message\\\":\\\"ue, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.53\\\\\\\", Port:8383, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.53\\\\\\\", Port:8081, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1008 00:07:55.166176 6981 services_controller.go:452] Built service openshift-marketplace/marketplace-operator-metrics per-node LB for network=default: []services.LB{}\\\\nI1008 00:07:55.166184 6981 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:960d98b2-dc64-4e93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1008 00:07:55.165765 6981 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, h\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:54Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-bs8w9_openshift-ovn-kubernetes(4f1af2ba-8907-4d14-afa1-ffde207c80f4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27fb2b1225e386dcd7fecaaa0aa54ba1fdfafabfde6c3824ae07315313067542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bs8w9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:56Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:56 crc kubenswrapper[4967]: I1008 00:07:56.762063 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-v5n7t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abe7ae2d-7e63-4249-96e1-4000e9628ccc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52575e853ff6727dde7c89e4add4362be126127db27191c2a7bf2871efa719d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpsjd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:06Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-v5n7t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:56Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:56 crc kubenswrapper[4967]: I1008 00:07:56.781160 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94e63bc5641ce80d507f76a19b203f78c42368a7a4105f3fd09e19af21a1b735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:56Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:56 crc kubenswrapper[4967]: I1008 00:07:56.785056 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:56 crc kubenswrapper[4967]: I1008 00:07:56.785126 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:56 crc kubenswrapper[4967]: I1008 00:07:56.785142 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:56 crc kubenswrapper[4967]: I1008 00:07:56.785160 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:56 crc kubenswrapper[4967]: I1008 00:07:56.785171 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:56Z","lastTransitionTime":"2025-10-08T00:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:56 crc kubenswrapper[4967]: I1008 00:07:56.796359 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e98666fd3152bab57d3723e7da72aae2703b79f7ca73eb3c3e5f0e062a82592\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cf1cbe5548146fe359aafa351864125ec6b8bdd7b796aa4d311f6f3a16ad212\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:56Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:56 crc kubenswrapper[4967]: I1008 00:07:56.807639 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9tdbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41264a12-9486-4fee-a08c-607b590e5a70\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://806e85df796ae300d279f73a0cc4433a5bb1f7620ee8ce33c8dc4674c4ce0596\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5blv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9tdbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:56Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:56 crc kubenswrapper[4967]: I1008 00:07:56.818446 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2h7z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59f68771-e25d-4367-87ee-68a31d767882\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7r2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7r2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:18Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2h7z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:56Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:56 crc kubenswrapper[4967]: I1008 00:07:56.833620 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6ea2440-b2ae-481e-9442-0b3dda9688ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37fd5dcdfb864320d4d17729ab48c6da505fd867f384ad70a6dcf5bed78a3978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c7893dfaa0370f159d41c30cd869bb18550a819e8c3d20dd6f57dcbcc697a57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c430979d298477b4394c3567bb3800ab7dea874cee8fa41a6a50a19a7e3a0133\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://426e86ab409776ff0fa9505dc0d8ec36f27a5bf21d3aedafb19d3935bdba7529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://426e86ab409776ff0fa9505dc0d8ec36f27a5bf21d3aedafb19d3935bdba7529\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:56Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:56 crc kubenswrapper[4967]: I1008 00:07:56.849348 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:56Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:56 crc kubenswrapper[4967]: I1008 00:07:56.874238 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-npz2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c71501b1-aa9e-420e-b1bd-7e15365e7adc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9010a5d6165a64f986f9101dbbbcaf8a7d3fa5af85c14be4bf3aa8ed8536ad4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82dfe6c420d804b2d7daef4b5cfe3f87b77b369287c4d65553fa93f5f24f589d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T00:07:51Z\\\",\\\"message\\\":\\\"2025-10-08T00:07:06+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_b0e5397b-daec-448a-8e92-e00fdfae8aea\\\\n2025-10-08T00:07:06+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_b0e5397b-daec-448a-8e92-e00fdfae8aea to /host/opt/cni/bin/\\\\n2025-10-08T00:07:06Z [verbose] multus-daemon started\\\\n2025-10-08T00:07:06Z [verbose] Readiness Indicator file check\\\\n2025-10-08T00:07:51Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gblk2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-npz2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:56Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:56 crc kubenswrapper[4967]: I1008 00:07:56.887654 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:56 crc kubenswrapper[4967]: I1008 00:07:56.887694 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:56 crc kubenswrapper[4967]: I1008 00:07:56.887705 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:56 crc kubenswrapper[4967]: I1008 00:07:56.887721 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:56 crc kubenswrapper[4967]: I1008 00:07:56.887733 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:56Z","lastTransitionTime":"2025-10-08T00:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:56 crc kubenswrapper[4967]: I1008 00:07:56.895785 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9mrr7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9299bc4f-f8d3-4baf-9885-39573982fd64\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://994beafece9dead0d85ae74dc453a142460705fd66b52ba22f2fa799cf27da08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb49c8476b482416588c4d403016940730064a619c0641eb029c947a4bab1215\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb49c8476b482416588c4d403016940730064a619c0641eb029c947a4bab1215\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b7b63962dcbf477f1fa4d71826a4aa3358920a2c1b2f5e48716411b5bdca049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b7b63962dcbf477f1fa4d71826a4aa3358920a2c1b2f5e48716411b5bdca049\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1e770dc737173fd4f3220ed6c52b422ccbd829722386a37029e1c3e10d80e1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1e770dc737173fd4f3220ed6c52b422ccbd829722386a37029e1c3e10d80e1a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e357d970878c996f219a79f1876b363176c9847b9575f33090f62bc299529a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e357d970878c996f219a79f1876b363176c9847b9575f33090f62bc299529a95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f098966bfad49d06278eaa72be042596d488115e597737d9a0a0cfc78f99f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49f098966bfad49d06278eaa72be042596d488115e597737d9a0a0cfc78f99f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1d53876b7b863c612d597fd716c3260e8857b6542802073bb730f6756071b46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c1d53876b7b863c612d597fd716c3260e8857b6542802073bb730f6756071b46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9mrr7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:56Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:56 crc kubenswrapper[4967]: I1008 00:07:56.908977 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kpmlg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f975d47d-42a6-422e-80e2-d1ee282817fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a317ae6640eb38b3aee0f9738cca890e8ab3d92075b620cd312f40a0628d46c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qznqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ef61854aea40924c5372b0553cd0b7e8faa39a1d42fec5b65cf867c368f72c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qznqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-kpmlg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:56Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:56 crc kubenswrapper[4967]: I1008 00:07:56.923649 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58c4704-a88a-43b4-b944-4de626f13f7e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22f5dabb39da0a7bd01db4dae1437b9a231a1fcc0dfe749a4a5a637a72cf33c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01a2b3dc785520c79eae998482e5a1347725c76584357073238fe60a619edb77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6aa6f5e81feceab3d1c7587152a70546b365e629a69267e4d3084324a1de764a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63c2b7c4b2791bd1f16888c58b878dbff1f87884e68c9949de914d0d3e1d39fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:56Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:56 crc kubenswrapper[4967]: I1008 00:07:56.940677 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:56Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:56 crc kubenswrapper[4967]: I1008 00:07:56.989792 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:56 crc kubenswrapper[4967]: I1008 00:07:56.990060 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:56 crc kubenswrapper[4967]: I1008 00:07:56.990146 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:56 crc kubenswrapper[4967]: I1008 00:07:56.990231 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:56 crc kubenswrapper[4967]: I1008 00:07:56.990319 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:56Z","lastTransitionTime":"2025-10-08T00:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:57 crc kubenswrapper[4967]: I1008 00:07:57.093447 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:57 crc kubenswrapper[4967]: I1008 00:07:57.093598 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:57 crc kubenswrapper[4967]: I1008 00:07:57.093620 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:57 crc kubenswrapper[4967]: I1008 00:07:57.093646 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:57 crc kubenswrapper[4967]: I1008 00:07:57.093663 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:57Z","lastTransitionTime":"2025-10-08T00:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:57 crc kubenswrapper[4967]: I1008 00:07:57.117655 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 00:07:57 crc kubenswrapper[4967]: I1008 00:07:57.117663 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 00:07:57 crc kubenswrapper[4967]: E1008 00:07:57.117889 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 00:07:57 crc kubenswrapper[4967]: E1008 00:07:57.118017 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 00:07:57 crc kubenswrapper[4967]: I1008 00:07:57.196327 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:57 crc kubenswrapper[4967]: I1008 00:07:57.196412 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:57 crc kubenswrapper[4967]: I1008 00:07:57.196490 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:57 crc kubenswrapper[4967]: I1008 00:07:57.196509 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:57 crc kubenswrapper[4967]: I1008 00:07:57.196522 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:57Z","lastTransitionTime":"2025-10-08T00:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:57 crc kubenswrapper[4967]: I1008 00:07:57.299553 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:57 crc kubenswrapper[4967]: I1008 00:07:57.299611 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:57 crc kubenswrapper[4967]: I1008 00:07:57.299628 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:57 crc kubenswrapper[4967]: I1008 00:07:57.299651 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:57 crc kubenswrapper[4967]: I1008 00:07:57.299701 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:57Z","lastTransitionTime":"2025-10-08T00:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:57 crc kubenswrapper[4967]: I1008 00:07:57.401712 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:57 crc kubenswrapper[4967]: I1008 00:07:57.401759 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:57 crc kubenswrapper[4967]: I1008 00:07:57.401772 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:57 crc kubenswrapper[4967]: I1008 00:07:57.401787 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:57 crc kubenswrapper[4967]: I1008 00:07:57.401800 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:57Z","lastTransitionTime":"2025-10-08T00:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:57 crc kubenswrapper[4967]: I1008 00:07:57.505013 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:57 crc kubenswrapper[4967]: I1008 00:07:57.505783 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:57 crc kubenswrapper[4967]: I1008 00:07:57.505918 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:57 crc kubenswrapper[4967]: I1008 00:07:57.506033 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:57 crc kubenswrapper[4967]: I1008 00:07:57.506158 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:57Z","lastTransitionTime":"2025-10-08T00:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:57 crc kubenswrapper[4967]: I1008 00:07:57.608185 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:57 crc kubenswrapper[4967]: I1008 00:07:57.608227 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:57 crc kubenswrapper[4967]: I1008 00:07:57.608237 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:57 crc kubenswrapper[4967]: I1008 00:07:57.608253 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:57 crc kubenswrapper[4967]: I1008 00:07:57.608266 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:57Z","lastTransitionTime":"2025-10-08T00:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:57 crc kubenswrapper[4967]: I1008 00:07:57.710336 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:57 crc kubenswrapper[4967]: I1008 00:07:57.710389 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:57 crc kubenswrapper[4967]: I1008 00:07:57.710402 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:57 crc kubenswrapper[4967]: I1008 00:07:57.710420 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:57 crc kubenswrapper[4967]: I1008 00:07:57.710433 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:57Z","lastTransitionTime":"2025-10-08T00:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:57 crc kubenswrapper[4967]: I1008 00:07:57.812824 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:57 crc kubenswrapper[4967]: I1008 00:07:57.812873 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:57 crc kubenswrapper[4967]: I1008 00:07:57.812889 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:57 crc kubenswrapper[4967]: I1008 00:07:57.812908 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:57 crc kubenswrapper[4967]: I1008 00:07:57.812921 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:57Z","lastTransitionTime":"2025-10-08T00:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:57 crc kubenswrapper[4967]: I1008 00:07:57.915306 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:57 crc kubenswrapper[4967]: I1008 00:07:57.915347 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:57 crc kubenswrapper[4967]: I1008 00:07:57.915357 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:57 crc kubenswrapper[4967]: I1008 00:07:57.915371 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:57 crc kubenswrapper[4967]: I1008 00:07:57.915382 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:57Z","lastTransitionTime":"2025-10-08T00:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:58 crc kubenswrapper[4967]: I1008 00:07:58.018031 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:58 crc kubenswrapper[4967]: I1008 00:07:58.018081 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:58 crc kubenswrapper[4967]: I1008 00:07:58.018094 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:58 crc kubenswrapper[4967]: I1008 00:07:58.018112 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:58 crc kubenswrapper[4967]: I1008 00:07:58.018124 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:58Z","lastTransitionTime":"2025-10-08T00:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:58 crc kubenswrapper[4967]: I1008 00:07:58.117373 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2h7z8" Oct 08 00:07:58 crc kubenswrapper[4967]: I1008 00:07:58.117404 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 00:07:58 crc kubenswrapper[4967]: E1008 00:07:58.117529 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2h7z8" podUID="59f68771-e25d-4367-87ee-68a31d767882" Oct 08 00:07:58 crc kubenswrapper[4967]: E1008 00:07:58.117644 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 00:07:58 crc kubenswrapper[4967]: I1008 00:07:58.120012 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:58 crc kubenswrapper[4967]: I1008 00:07:58.120044 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:58 crc kubenswrapper[4967]: I1008 00:07:58.120054 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:58 crc kubenswrapper[4967]: I1008 00:07:58.120069 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:58 crc kubenswrapper[4967]: I1008 00:07:58.120081 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:58Z","lastTransitionTime":"2025-10-08T00:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:58 crc kubenswrapper[4967]: I1008 00:07:58.221973 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:58 crc kubenswrapper[4967]: I1008 00:07:58.222015 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:58 crc kubenswrapper[4967]: I1008 00:07:58.222024 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:58 crc kubenswrapper[4967]: I1008 00:07:58.222041 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:58 crc kubenswrapper[4967]: I1008 00:07:58.222050 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:58Z","lastTransitionTime":"2025-10-08T00:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:58 crc kubenswrapper[4967]: I1008 00:07:58.324238 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:58 crc kubenswrapper[4967]: I1008 00:07:58.324276 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:58 crc kubenswrapper[4967]: I1008 00:07:58.324285 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:58 crc kubenswrapper[4967]: I1008 00:07:58.324302 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:58 crc kubenswrapper[4967]: I1008 00:07:58.324312 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:58Z","lastTransitionTime":"2025-10-08T00:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:58 crc kubenswrapper[4967]: I1008 00:07:58.427016 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:58 crc kubenswrapper[4967]: I1008 00:07:58.427081 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:58 crc kubenswrapper[4967]: I1008 00:07:58.427097 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:58 crc kubenswrapper[4967]: I1008 00:07:58.427115 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:58 crc kubenswrapper[4967]: I1008 00:07:58.427127 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:58Z","lastTransitionTime":"2025-10-08T00:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:58 crc kubenswrapper[4967]: I1008 00:07:58.529218 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:58 crc kubenswrapper[4967]: I1008 00:07:58.529292 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:58 crc kubenswrapper[4967]: I1008 00:07:58.529315 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:58 crc kubenswrapper[4967]: I1008 00:07:58.529337 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:58 crc kubenswrapper[4967]: I1008 00:07:58.529355 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:58Z","lastTransitionTime":"2025-10-08T00:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:58 crc kubenswrapper[4967]: I1008 00:07:58.631865 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:58 crc kubenswrapper[4967]: I1008 00:07:58.631918 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:58 crc kubenswrapper[4967]: I1008 00:07:58.631933 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:58 crc kubenswrapper[4967]: I1008 00:07:58.631954 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:58 crc kubenswrapper[4967]: I1008 00:07:58.631968 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:58Z","lastTransitionTime":"2025-10-08T00:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:58 crc kubenswrapper[4967]: I1008 00:07:58.735105 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:58 crc kubenswrapper[4967]: I1008 00:07:58.735148 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:58 crc kubenswrapper[4967]: I1008 00:07:58.735158 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:58 crc kubenswrapper[4967]: I1008 00:07:58.735175 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:58 crc kubenswrapper[4967]: I1008 00:07:58.735184 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:58Z","lastTransitionTime":"2025-10-08T00:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:58 crc kubenswrapper[4967]: I1008 00:07:58.837215 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:58 crc kubenswrapper[4967]: I1008 00:07:58.837272 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:58 crc kubenswrapper[4967]: I1008 00:07:58.837282 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:58 crc kubenswrapper[4967]: I1008 00:07:58.837294 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:58 crc kubenswrapper[4967]: I1008 00:07:58.837302 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:58Z","lastTransitionTime":"2025-10-08T00:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:58 crc kubenswrapper[4967]: I1008 00:07:58.939060 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:58 crc kubenswrapper[4967]: I1008 00:07:58.939116 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:58 crc kubenswrapper[4967]: I1008 00:07:58.939126 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:58 crc kubenswrapper[4967]: I1008 00:07:58.939139 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:58 crc kubenswrapper[4967]: I1008 00:07:58.939147 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:58Z","lastTransitionTime":"2025-10-08T00:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:59 crc kubenswrapper[4967]: I1008 00:07:59.040982 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:59 crc kubenswrapper[4967]: I1008 00:07:59.041045 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:59 crc kubenswrapper[4967]: I1008 00:07:59.041055 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:59 crc kubenswrapper[4967]: I1008 00:07:59.041066 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:59 crc kubenswrapper[4967]: I1008 00:07:59.041074 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:59Z","lastTransitionTime":"2025-10-08T00:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:59 crc kubenswrapper[4967]: I1008 00:07:59.117764 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 00:07:59 crc kubenswrapper[4967]: I1008 00:07:59.117805 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 00:07:59 crc kubenswrapper[4967]: E1008 00:07:59.117898 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 00:07:59 crc kubenswrapper[4967]: E1008 00:07:59.118007 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 00:07:59 crc kubenswrapper[4967]: I1008 00:07:59.143840 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:59 crc kubenswrapper[4967]: I1008 00:07:59.143898 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:59 crc kubenswrapper[4967]: I1008 00:07:59.143911 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:59 crc kubenswrapper[4967]: I1008 00:07:59.143929 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:59 crc kubenswrapper[4967]: I1008 00:07:59.143943 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:59Z","lastTransitionTime":"2025-10-08T00:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:59 crc kubenswrapper[4967]: I1008 00:07:59.246898 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:59 crc kubenswrapper[4967]: I1008 00:07:59.246945 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:59 crc kubenswrapper[4967]: I1008 00:07:59.246955 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:59 crc kubenswrapper[4967]: I1008 00:07:59.246971 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:59 crc kubenswrapper[4967]: I1008 00:07:59.246980 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:59Z","lastTransitionTime":"2025-10-08T00:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:59 crc kubenswrapper[4967]: I1008 00:07:59.350026 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:59 crc kubenswrapper[4967]: I1008 00:07:59.350072 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:59 crc kubenswrapper[4967]: I1008 00:07:59.350084 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:59 crc kubenswrapper[4967]: I1008 00:07:59.350100 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:59 crc kubenswrapper[4967]: I1008 00:07:59.350111 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:59Z","lastTransitionTime":"2025-10-08T00:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:59 crc kubenswrapper[4967]: I1008 00:07:59.454851 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:59 crc kubenswrapper[4967]: I1008 00:07:59.454923 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:59 crc kubenswrapper[4967]: I1008 00:07:59.454944 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:59 crc kubenswrapper[4967]: I1008 00:07:59.454971 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:59 crc kubenswrapper[4967]: I1008 00:07:59.454988 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:59Z","lastTransitionTime":"2025-10-08T00:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:59 crc kubenswrapper[4967]: I1008 00:07:59.557693 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:59 crc kubenswrapper[4967]: I1008 00:07:59.557741 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:59 crc kubenswrapper[4967]: I1008 00:07:59.557749 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:59 crc kubenswrapper[4967]: I1008 00:07:59.557764 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:59 crc kubenswrapper[4967]: I1008 00:07:59.557774 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:59Z","lastTransitionTime":"2025-10-08T00:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:59 crc kubenswrapper[4967]: I1008 00:07:59.660965 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:59 crc kubenswrapper[4967]: I1008 00:07:59.661079 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:59 crc kubenswrapper[4967]: I1008 00:07:59.661106 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:59 crc kubenswrapper[4967]: I1008 00:07:59.661134 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:59 crc kubenswrapper[4967]: I1008 00:07:59.661159 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:59Z","lastTransitionTime":"2025-10-08T00:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:59 crc kubenswrapper[4967]: I1008 00:07:59.764344 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:59 crc kubenswrapper[4967]: I1008 00:07:59.764381 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:59 crc kubenswrapper[4967]: I1008 00:07:59.764392 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:59 crc kubenswrapper[4967]: I1008 00:07:59.764409 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:59 crc kubenswrapper[4967]: I1008 00:07:59.764420 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:59Z","lastTransitionTime":"2025-10-08T00:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:59 crc kubenswrapper[4967]: I1008 00:07:59.868045 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:59 crc kubenswrapper[4967]: I1008 00:07:59.868131 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:59 crc kubenswrapper[4967]: I1008 00:07:59.868155 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:59 crc kubenswrapper[4967]: I1008 00:07:59.868185 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:59 crc kubenswrapper[4967]: I1008 00:07:59.868207 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:59Z","lastTransitionTime":"2025-10-08T00:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:59 crc kubenswrapper[4967]: I1008 00:07:59.880218 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:59 crc kubenswrapper[4967]: I1008 00:07:59.880269 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:59 crc kubenswrapper[4967]: I1008 00:07:59.880286 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:59 crc kubenswrapper[4967]: I1008 00:07:59.880309 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:59 crc kubenswrapper[4967]: I1008 00:07:59.880375 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:59Z","lastTransitionTime":"2025-10-08T00:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:59 crc kubenswrapper[4967]: E1008 00:07:59.897325 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5cddad4c-22d1-42d7-a373-87b47c1cc97b\\\",\\\"systemUUID\\\":\\\"b243ca1c-20cd-4e6f-96b2-52fce74dd2d4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:59Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:59 crc kubenswrapper[4967]: I1008 00:07:59.902397 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:59 crc kubenswrapper[4967]: I1008 00:07:59.902438 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:59 crc kubenswrapper[4967]: I1008 00:07:59.902455 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:59 crc kubenswrapper[4967]: I1008 00:07:59.902477 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:59 crc kubenswrapper[4967]: I1008 00:07:59.902491 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:59Z","lastTransitionTime":"2025-10-08T00:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:59 crc kubenswrapper[4967]: E1008 00:07:59.918939 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5cddad4c-22d1-42d7-a373-87b47c1cc97b\\\",\\\"systemUUID\\\":\\\"b243ca1c-20cd-4e6f-96b2-52fce74dd2d4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:59Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:59 crc kubenswrapper[4967]: I1008 00:07:59.923209 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:59 crc kubenswrapper[4967]: I1008 00:07:59.923269 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:59 crc kubenswrapper[4967]: I1008 00:07:59.923286 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:59 crc kubenswrapper[4967]: I1008 00:07:59.923311 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:59 crc kubenswrapper[4967]: I1008 00:07:59.923324 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:59Z","lastTransitionTime":"2025-10-08T00:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:59 crc kubenswrapper[4967]: E1008 00:07:59.936248 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5cddad4c-22d1-42d7-a373-87b47c1cc97b\\\",\\\"systemUUID\\\":\\\"b243ca1c-20cd-4e6f-96b2-52fce74dd2d4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:59Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:59 crc kubenswrapper[4967]: I1008 00:07:59.940476 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:59 crc kubenswrapper[4967]: I1008 00:07:59.940529 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:59 crc kubenswrapper[4967]: I1008 00:07:59.940544 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:59 crc kubenswrapper[4967]: I1008 00:07:59.940563 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:59 crc kubenswrapper[4967]: I1008 00:07:59.940575 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:59Z","lastTransitionTime":"2025-10-08T00:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:59 crc kubenswrapper[4967]: E1008 00:07:59.953654 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5cddad4c-22d1-42d7-a373-87b47c1cc97b\\\",\\\"systemUUID\\\":\\\"b243ca1c-20cd-4e6f-96b2-52fce74dd2d4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:59Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:59 crc kubenswrapper[4967]: I1008 00:07:59.957388 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:59 crc kubenswrapper[4967]: I1008 00:07:59.957470 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:59 crc kubenswrapper[4967]: I1008 00:07:59.957485 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:59 crc kubenswrapper[4967]: I1008 00:07:59.957504 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:59 crc kubenswrapper[4967]: I1008 00:07:59.957514 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:59Z","lastTransitionTime":"2025-10-08T00:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:07:59 crc kubenswrapper[4967]: E1008 00:07:59.972197 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T00:07:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5cddad4c-22d1-42d7-a373-87b47c1cc97b\\\",\\\"systemUUID\\\":\\\"b243ca1c-20cd-4e6f-96b2-52fce74dd2d4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:07:59Z is after 2025-08-24T17:21:41Z" Oct 08 00:07:59 crc kubenswrapper[4967]: E1008 00:07:59.972426 4967 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 08 00:07:59 crc kubenswrapper[4967]: I1008 00:07:59.974481 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:07:59 crc kubenswrapper[4967]: I1008 00:07:59.974552 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:07:59 crc kubenswrapper[4967]: I1008 00:07:59.974565 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:07:59 crc kubenswrapper[4967]: I1008 00:07:59.974589 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:07:59 crc kubenswrapper[4967]: I1008 00:07:59.974602 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:07:59Z","lastTransitionTime":"2025-10-08T00:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:00 crc kubenswrapper[4967]: I1008 00:08:00.077718 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:00 crc kubenswrapper[4967]: I1008 00:08:00.077783 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:00 crc kubenswrapper[4967]: I1008 00:08:00.077792 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:00 crc kubenswrapper[4967]: I1008 00:08:00.077805 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:00 crc kubenswrapper[4967]: I1008 00:08:00.077814 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:00Z","lastTransitionTime":"2025-10-08T00:08:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:00 crc kubenswrapper[4967]: I1008 00:08:00.117397 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 00:08:00 crc kubenswrapper[4967]: I1008 00:08:00.117518 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2h7z8" Oct 08 00:08:00 crc kubenswrapper[4967]: E1008 00:08:00.117716 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 00:08:00 crc kubenswrapper[4967]: E1008 00:08:00.117987 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2h7z8" podUID="59f68771-e25d-4367-87ee-68a31d767882" Oct 08 00:08:00 crc kubenswrapper[4967]: I1008 00:08:00.179904 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:00 crc kubenswrapper[4967]: I1008 00:08:00.179957 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:00 crc kubenswrapper[4967]: I1008 00:08:00.179972 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:00 crc kubenswrapper[4967]: I1008 00:08:00.179994 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:00 crc kubenswrapper[4967]: I1008 00:08:00.180006 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:00Z","lastTransitionTime":"2025-10-08T00:08:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:00 crc kubenswrapper[4967]: I1008 00:08:00.283338 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:00 crc kubenswrapper[4967]: I1008 00:08:00.283379 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:00 crc kubenswrapper[4967]: I1008 00:08:00.283391 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:00 crc kubenswrapper[4967]: I1008 00:08:00.283407 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:00 crc kubenswrapper[4967]: I1008 00:08:00.283440 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:00Z","lastTransitionTime":"2025-10-08T00:08:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:00 crc kubenswrapper[4967]: I1008 00:08:00.386034 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:00 crc kubenswrapper[4967]: I1008 00:08:00.386082 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:00 crc kubenswrapper[4967]: I1008 00:08:00.386095 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:00 crc kubenswrapper[4967]: I1008 00:08:00.386109 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:00 crc kubenswrapper[4967]: I1008 00:08:00.386119 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:00Z","lastTransitionTime":"2025-10-08T00:08:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:00 crc kubenswrapper[4967]: I1008 00:08:00.489273 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:00 crc kubenswrapper[4967]: I1008 00:08:00.489386 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:00 crc kubenswrapper[4967]: I1008 00:08:00.489407 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:00 crc kubenswrapper[4967]: I1008 00:08:00.489432 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:00 crc kubenswrapper[4967]: I1008 00:08:00.489481 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:00Z","lastTransitionTime":"2025-10-08T00:08:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:00 crc kubenswrapper[4967]: I1008 00:08:00.592253 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:00 crc kubenswrapper[4967]: I1008 00:08:00.592298 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:00 crc kubenswrapper[4967]: I1008 00:08:00.592314 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:00 crc kubenswrapper[4967]: I1008 00:08:00.592336 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:00 crc kubenswrapper[4967]: I1008 00:08:00.592357 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:00Z","lastTransitionTime":"2025-10-08T00:08:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:00 crc kubenswrapper[4967]: I1008 00:08:00.696026 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:00 crc kubenswrapper[4967]: I1008 00:08:00.696103 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:00 crc kubenswrapper[4967]: I1008 00:08:00.696127 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:00 crc kubenswrapper[4967]: I1008 00:08:00.696156 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:00 crc kubenswrapper[4967]: I1008 00:08:00.696176 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:00Z","lastTransitionTime":"2025-10-08T00:08:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:00 crc kubenswrapper[4967]: I1008 00:08:00.798929 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:00 crc kubenswrapper[4967]: I1008 00:08:00.799235 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:00 crc kubenswrapper[4967]: I1008 00:08:00.799342 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:00 crc kubenswrapper[4967]: I1008 00:08:00.799434 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:00 crc kubenswrapper[4967]: I1008 00:08:00.799515 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:00Z","lastTransitionTime":"2025-10-08T00:08:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:00 crc kubenswrapper[4967]: I1008 00:08:00.902138 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:00 crc kubenswrapper[4967]: I1008 00:08:00.902186 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:00 crc kubenswrapper[4967]: I1008 00:08:00.902198 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:00 crc kubenswrapper[4967]: I1008 00:08:00.902216 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:00 crc kubenswrapper[4967]: I1008 00:08:00.902228 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:00Z","lastTransitionTime":"2025-10-08T00:08:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:01 crc kubenswrapper[4967]: I1008 00:08:01.004293 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:01 crc kubenswrapper[4967]: I1008 00:08:01.004340 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:01 crc kubenswrapper[4967]: I1008 00:08:01.004355 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:01 crc kubenswrapper[4967]: I1008 00:08:01.004375 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:01 crc kubenswrapper[4967]: I1008 00:08:01.004388 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:01Z","lastTransitionTime":"2025-10-08T00:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:01 crc kubenswrapper[4967]: I1008 00:08:01.107401 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:01 crc kubenswrapper[4967]: I1008 00:08:01.107473 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:01 crc kubenswrapper[4967]: I1008 00:08:01.107496 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:01 crc kubenswrapper[4967]: I1008 00:08:01.107522 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:01 crc kubenswrapper[4967]: I1008 00:08:01.107539 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:01Z","lastTransitionTime":"2025-10-08T00:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:01 crc kubenswrapper[4967]: I1008 00:08:01.117720 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 00:08:01 crc kubenswrapper[4967]: I1008 00:08:01.117741 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 00:08:01 crc kubenswrapper[4967]: E1008 00:08:01.117841 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 00:08:01 crc kubenswrapper[4967]: E1008 00:08:01.118034 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 00:08:01 crc kubenswrapper[4967]: I1008 00:08:01.210848 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:01 crc kubenswrapper[4967]: I1008 00:08:01.210944 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:01 crc kubenswrapper[4967]: I1008 00:08:01.210971 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:01 crc kubenswrapper[4967]: I1008 00:08:01.211123 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:01 crc kubenswrapper[4967]: I1008 00:08:01.211149 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:01Z","lastTransitionTime":"2025-10-08T00:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:01 crc kubenswrapper[4967]: I1008 00:08:01.314278 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:01 crc kubenswrapper[4967]: I1008 00:08:01.314312 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:01 crc kubenswrapper[4967]: I1008 00:08:01.314320 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:01 crc kubenswrapper[4967]: I1008 00:08:01.314334 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:01 crc kubenswrapper[4967]: I1008 00:08:01.314343 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:01Z","lastTransitionTime":"2025-10-08T00:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:01 crc kubenswrapper[4967]: I1008 00:08:01.416717 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:01 crc kubenswrapper[4967]: I1008 00:08:01.416773 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:01 crc kubenswrapper[4967]: I1008 00:08:01.416787 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:01 crc kubenswrapper[4967]: I1008 00:08:01.416806 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:01 crc kubenswrapper[4967]: I1008 00:08:01.416818 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:01Z","lastTransitionTime":"2025-10-08T00:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:01 crc kubenswrapper[4967]: I1008 00:08:01.519894 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:01 crc kubenswrapper[4967]: I1008 00:08:01.519975 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:01 crc kubenswrapper[4967]: I1008 00:08:01.519995 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:01 crc kubenswrapper[4967]: I1008 00:08:01.520022 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:01 crc kubenswrapper[4967]: I1008 00:08:01.520040 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:01Z","lastTransitionTime":"2025-10-08T00:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:01 crc kubenswrapper[4967]: I1008 00:08:01.622195 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:01 crc kubenswrapper[4967]: I1008 00:08:01.622889 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:01 crc kubenswrapper[4967]: I1008 00:08:01.623022 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:01 crc kubenswrapper[4967]: I1008 00:08:01.623132 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:01 crc kubenswrapper[4967]: I1008 00:08:01.623224 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:01Z","lastTransitionTime":"2025-10-08T00:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:01 crc kubenswrapper[4967]: I1008 00:08:01.725065 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:01 crc kubenswrapper[4967]: I1008 00:08:01.725321 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:01 crc kubenswrapper[4967]: I1008 00:08:01.725402 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:01 crc kubenswrapper[4967]: I1008 00:08:01.725479 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:01 crc kubenswrapper[4967]: I1008 00:08:01.725553 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:01Z","lastTransitionTime":"2025-10-08T00:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:01 crc kubenswrapper[4967]: I1008 00:08:01.827786 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:01 crc kubenswrapper[4967]: I1008 00:08:01.827817 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:01 crc kubenswrapper[4967]: I1008 00:08:01.827828 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:01 crc kubenswrapper[4967]: I1008 00:08:01.827840 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:01 crc kubenswrapper[4967]: I1008 00:08:01.827848 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:01Z","lastTransitionTime":"2025-10-08T00:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:01 crc kubenswrapper[4967]: I1008 00:08:01.932204 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:01 crc kubenswrapper[4967]: I1008 00:08:01.932238 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:01 crc kubenswrapper[4967]: I1008 00:08:01.932246 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:01 crc kubenswrapper[4967]: I1008 00:08:01.932262 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:01 crc kubenswrapper[4967]: I1008 00:08:01.932271 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:01Z","lastTransitionTime":"2025-10-08T00:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:02 crc kubenswrapper[4967]: I1008 00:08:02.034857 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:02 crc kubenswrapper[4967]: I1008 00:08:02.035273 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:02 crc kubenswrapper[4967]: I1008 00:08:02.035431 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:02 crc kubenswrapper[4967]: I1008 00:08:02.035584 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:02 crc kubenswrapper[4967]: I1008 00:08:02.035749 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:02Z","lastTransitionTime":"2025-10-08T00:08:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:02 crc kubenswrapper[4967]: I1008 00:08:02.117542 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2h7z8" Oct 08 00:08:02 crc kubenswrapper[4967]: I1008 00:08:02.117581 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 00:08:02 crc kubenswrapper[4967]: E1008 00:08:02.117744 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2h7z8" podUID="59f68771-e25d-4367-87ee-68a31d767882" Oct 08 00:08:02 crc kubenswrapper[4967]: E1008 00:08:02.117917 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 00:08:02 crc kubenswrapper[4967]: I1008 00:08:02.131248 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0fa4b747-ab6e-4a11-9903-ef2dcb9668c4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fbadcdade274b6d9be940f22bc949e3c55cadf95dcad43c8cb7a1085c0e6708c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d0a0b2ef4a98ee828688081a74f4b7ecc31c8fae415d7257a560ecd43960ce3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b96b53b1b283025b3d078a3f505d469d1872ad4bb8ab8e612c36c70b0b00a920\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef86a2422616e1e0d5c3afe844818c2bdf2150f941d1aefc506995f58587b2bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1400e110997c64fa33d6b8672c7e91cab5f782952ddf8e1dea2a882689026e7f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1008 00:06:56.760088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 00:06:56.761210 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-369805205/tls.crt::/tmp/serving-cert-369805205/tls.key\\\\\\\"\\\\nI1008 00:07:02.321448 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1008 00:07:02.325932 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1008 00:07:02.325980 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1008 00:07:02.326025 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1008 00:07:02.326042 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1008 00:07:02.336203 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1008 00:07:02.336240 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 00:07:02.336245 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1008 00:07:02.336250 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1008 00:07:02.336253 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1008 00:07:02.336256 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1008 00:07:02.336258 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1008 00:07:02.336430 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1008 00:07:02.340774 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:46Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ca28d8db19036e47a7ef3dadb70941f6c6cf5735e7703977d036185c46fce40\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b5b222ad84c31d8746dfe03124c7287b401deb3ca2c62ce2b4fbd40e638cee0e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5b222ad84c31d8746dfe03124c7287b401deb3ca2c62ce2b4fbd40e638cee0e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:08:02Z is after 2025-08-24T17:21:41Z" Oct 08 00:08:02 crc kubenswrapper[4967]: I1008 00:08:02.138582 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:02 crc kubenswrapper[4967]: I1008 00:08:02.139069 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:02 crc kubenswrapper[4967]: I1008 00:08:02.139344 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:02 crc kubenswrapper[4967]: I1008 00:08:02.139649 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:02 crc kubenswrapper[4967]: I1008 00:08:02.139968 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:02Z","lastTransitionTime":"2025-10-08T00:08:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:02 crc kubenswrapper[4967]: I1008 00:08:02.143385 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dc98bc5e-fbae-4c67-a5c3-9738b3e966cd\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acfae51bf8bee1b379a3859b7fb98ad0df3bed5d49dc0abe82b917512955c009\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ae14677731c53e87871129d2140a34b18f0a48555309b652ebd490996fb1a723\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae14677731c53e87871129d2140a34b18f0a48555309b652ebd490996fb1a723\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:08:02Z is after 2025-08-24T17:21:41Z" Oct 08 00:08:02 crc kubenswrapper[4967]: I1008 00:08:02.177691 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:08:02Z is after 2025-08-24T17:21:41Z" Oct 08 00:08:02 crc kubenswrapper[4967]: I1008 00:08:02.210518 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dbfcdc882e2be1cb9cf666c9e7775671078f7dbbfe8421dd15db737b89920a36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:08:02Z is after 2025-08-24T17:21:41Z" Oct 08 00:08:02 crc kubenswrapper[4967]: I1008 00:08:02.229896 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9461162a-c27d-465e-9ca7-2d3936877cd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9778d78ddc7325abd4d77bc5e71b0b114f50be5dda4a438c7cb6b05640c8a07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z9tp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28760af04ef321d0f91b2bce3ad074d44b63dabf676cc37f3be6aa8726341a2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-z9tp4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xvdtp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:08:02Z is after 2025-08-24T17:21:41Z" Oct 08 00:08:02 crc kubenswrapper[4967]: I1008 00:08:02.241978 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:02 crc kubenswrapper[4967]: I1008 00:08:02.242313 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:02 crc kubenswrapper[4967]: I1008 00:08:02.242426 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:02 crc kubenswrapper[4967]: I1008 00:08:02.242538 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:02 crc kubenswrapper[4967]: I1008 00:08:02.242626 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:02Z","lastTransitionTime":"2025-10-08T00:08:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:02 crc kubenswrapper[4967]: I1008 00:08:02.242369 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94e63bc5641ce80d507f76a19b203f78c42368a7a4105f3fd09e19af21a1b735\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:08:02Z is after 2025-08-24T17:21:41Z" Oct 08 00:08:02 crc kubenswrapper[4967]: I1008 00:08:02.256404 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5e98666fd3152bab57d3723e7da72aae2703b79f7ca73eb3c3e5f0e062a82592\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cf1cbe5548146fe359aafa351864125ec6b8bdd7b796aa4d311f6f3a16ad212\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:08:02Z is after 2025-08-24T17:21:41Z" Oct 08 00:08:02 crc kubenswrapper[4967]: I1008 00:08:02.268103 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9tdbb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41264a12-9486-4fee-a08c-607b590e5a70\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://806e85df796ae300d279f73a0cc4433a5bb1f7620ee8ce33c8dc4674c4ce0596\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s5blv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9tdbb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:08:02Z is after 2025-08-24T17:21:41Z" Oct 08 00:08:02 crc kubenswrapper[4967]: I1008 00:08:02.286879 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4f1af2ba-8907-4d14-afa1-ffde207c80f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4535cffb8c9d64e662e282ecb44f6817b4604866f0bb75015f3b888d5554f821\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b03b83ddff429dbfa8fe052901aaa86a39ff0dd75477ffa8a6b6f4198f5819ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c9fa0559eb35ba67e11e36069ac8d0b2fec76399788eef668a08ad4353df3ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3312263e12bd42fbed0476ab5ace66b5497f62c9477865e130d67e3922e05cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e69dc603055aa221d44d11c049ddcc50cec568979911a310ed773a90341a499\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115f7a33f3dedae8f59d2f45f465f95f48b99c824e895ef7a0be9fe7a1b61d23\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f7d46ee97fd3c4286f45879fededea788052cf83ebeebcb539be452f2ae102d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f7d46ee97fd3c4286f45879fededea788052cf83ebeebcb539be452f2ae102d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T00:07:55Z\\\",\\\"message\\\":\\\"ue, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.53\\\\\\\", Port:8383, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}, services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.53\\\\\\\", Port:8081, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1008 00:07:55.166176 6981 services_controller.go:452] Built service openshift-marketplace/marketplace-operator-metrics per-node LB for network=default: []services.LB{}\\\\nI1008 00:07:55.166184 6981 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:960d98b2-dc64-4e93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1008 00:07:55.165765 6981 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, h\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:54Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-bs8w9_openshift-ovn-kubernetes(4f1af2ba-8907-4d14-afa1-ffde207c80f4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27fb2b1225e386dcd7fecaaa0aa54ba1fdfafabfde6c3824ae07315313067542\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qlrlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bs8w9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:08:02Z is after 2025-08-24T17:21:41Z" Oct 08 00:08:02 crc kubenswrapper[4967]: I1008 00:08:02.296628 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-v5n7t" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abe7ae2d-7e63-4249-96e1-4000e9628ccc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://52575e853ff6727dde7c89e4add4362be126127db27191c2a7bf2871efa719d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bpsjd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:06Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-v5n7t\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:08:02Z is after 2025-08-24T17:21:41Z" Oct 08 00:08:02 crc kubenswrapper[4967]: I1008 00:08:02.310130 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6ea2440-b2ae-481e-9442-0b3dda9688ef\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://37fd5dcdfb864320d4d17729ab48c6da505fd867f384ad70a6dcf5bed78a3978\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9c7893dfaa0370f159d41c30cd869bb18550a819e8c3d20dd6f57dcbcc697a57\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c430979d298477b4394c3567bb3800ab7dea874cee8fa41a6a50a19a7e3a0133\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://426e86ab409776ff0fa9505dc0d8ec36f27a5bf21d3aedafb19d3935bdba7529\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://426e86ab409776ff0fa9505dc0d8ec36f27a5bf21d3aedafb19d3935bdba7529\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:08:02Z is after 2025-08-24T17:21:41Z" Oct 08 00:08:02 crc kubenswrapper[4967]: I1008 00:08:02.320224 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:08:02Z is after 2025-08-24T17:21:41Z" Oct 08 00:08:02 crc kubenswrapper[4967]: I1008 00:08:02.331327 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-npz2s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c71501b1-aa9e-420e-b1bd-7e15365e7adc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9010a5d6165a64f986f9101dbbbcaf8a7d3fa5af85c14be4bf3aa8ed8536ad4e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82dfe6c420d804b2d7daef4b5cfe3f87b77b369287c4d65553fa93f5f24f589d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T00:07:51Z\\\",\\\"message\\\":\\\"2025-10-08T00:07:06+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_b0e5397b-daec-448a-8e92-e00fdfae8aea\\\\n2025-10-08T00:07:06+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_b0e5397b-daec-448a-8e92-e00fdfae8aea to /host/opt/cni/bin/\\\\n2025-10-08T00:07:06Z [verbose] multus-daemon started\\\\n2025-10-08T00:07:06Z [verbose] Readiness Indicator file check\\\\n2025-10-08T00:07:51Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gblk2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-npz2s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:08:02Z is after 2025-08-24T17:21:41Z" Oct 08 00:08:02 crc kubenswrapper[4967]: I1008 00:08:02.340676 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-2h7z8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59f68771-e25d-4367-87ee-68a31d767882\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7r2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c7r2h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:18Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-2h7z8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:08:02Z is after 2025-08-24T17:21:41Z" Oct 08 00:08:02 crc kubenswrapper[4967]: I1008 00:08:02.344301 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:02 crc kubenswrapper[4967]: I1008 00:08:02.344414 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:02 crc kubenswrapper[4967]: I1008 00:08:02.344494 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:02 crc kubenswrapper[4967]: I1008 00:08:02.344585 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:02 crc kubenswrapper[4967]: I1008 00:08:02.344675 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:02Z","lastTransitionTime":"2025-10-08T00:08:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:02 crc kubenswrapper[4967]: I1008 00:08:02.352634 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d58c4704-a88a-43b4-b944-4de626f13f7e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:06:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a22f5dabb39da0a7bd01db4dae1437b9a231a1fcc0dfe749a4a5a637a72cf33c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://01a2b3dc785520c79eae998482e5a1347725c76584357073238fe60a619edb77\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6aa6f5e81feceab3d1c7587152a70546b365e629a69267e4d3084324a1de764a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://63c2b7c4b2791bd1f16888c58b878dbff1f87884e68c9949de914d0d3e1d39fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:06:42Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:08:02Z is after 2025-08-24T17:21:41Z" Oct 08 00:08:02 crc kubenswrapper[4967]: I1008 00:08:02.362377 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:03Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:08:02Z is after 2025-08-24T17:21:41Z" Oct 08 00:08:02 crc kubenswrapper[4967]: I1008 00:08:02.374007 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-9mrr7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9299bc4f-f8d3-4baf-9885-39573982fd64\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://994beafece9dead0d85ae74dc453a142460705fd66b52ba22f2fa799cf27da08\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb49c8476b482416588c4d403016940730064a619c0641eb029c947a4bab1215\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fb49c8476b482416588c4d403016940730064a619c0641eb029c947a4bab1215\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b7b63962dcbf477f1fa4d71826a4aa3358920a2c1b2f5e48716411b5bdca049\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b7b63962dcbf477f1fa4d71826a4aa3358920a2c1b2f5e48716411b5bdca049\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1e770dc737173fd4f3220ed6c52b422ccbd829722386a37029e1c3e10d80e1a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1e770dc737173fd4f3220ed6c52b422ccbd829722386a37029e1c3e10d80e1a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e357d970878c996f219a79f1876b363176c9847b9575f33090f62bc299529a95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e357d970878c996f219a79f1876b363176c9847b9575f33090f62bc299529a95\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://49f098966bfad49d06278eaa72be042596d488115e597737d9a0a0cfc78f99f1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://49f098966bfad49d06278eaa72be042596d488115e597737d9a0a0cfc78f99f1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1d53876b7b863c612d597fd716c3260e8857b6542802073bb730f6756071b46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c1d53876b7b863c612d597fd716c3260e8857b6542802073bb730f6756071b46\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T00:07:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dbsgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:04Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-9mrr7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:08:02Z is after 2025-08-24T17:21:41Z" Oct 08 00:08:02 crc kubenswrapper[4967]: I1008 00:08:02.384896 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kpmlg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f975d47d-42a6-422e-80e2-d1ee282817fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T00:07:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a317ae6640eb38b3aee0f9738cca890e8ab3d92075b620cd312f40a0628d46c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qznqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ef61854aea40924c5372b0553cd0b7e8faa39a1d42fec5b65cf867c368f72c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T00:07:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qznqw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T00:07:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-kpmlg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T00:08:02Z is after 2025-08-24T17:21:41Z" Oct 08 00:08:02 crc kubenswrapper[4967]: I1008 00:08:02.446942 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:02 crc kubenswrapper[4967]: I1008 00:08:02.446996 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:02 crc kubenswrapper[4967]: I1008 00:08:02.447009 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:02 crc kubenswrapper[4967]: I1008 00:08:02.447024 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:02 crc kubenswrapper[4967]: I1008 00:08:02.447035 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:02Z","lastTransitionTime":"2025-10-08T00:08:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:02 crc kubenswrapper[4967]: I1008 00:08:02.550156 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:02 crc kubenswrapper[4967]: I1008 00:08:02.550202 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:02 crc kubenswrapper[4967]: I1008 00:08:02.550215 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:02 crc kubenswrapper[4967]: I1008 00:08:02.550233 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:02 crc kubenswrapper[4967]: I1008 00:08:02.550245 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:02Z","lastTransitionTime":"2025-10-08T00:08:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:02 crc kubenswrapper[4967]: I1008 00:08:02.662081 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:02 crc kubenswrapper[4967]: I1008 00:08:02.662193 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:02 crc kubenswrapper[4967]: I1008 00:08:02.662211 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:02 crc kubenswrapper[4967]: I1008 00:08:02.662235 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:02 crc kubenswrapper[4967]: I1008 00:08:02.662250 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:02Z","lastTransitionTime":"2025-10-08T00:08:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:02 crc kubenswrapper[4967]: I1008 00:08:02.764169 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:02 crc kubenswrapper[4967]: I1008 00:08:02.764227 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:02 crc kubenswrapper[4967]: I1008 00:08:02.764240 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:02 crc kubenswrapper[4967]: I1008 00:08:02.764256 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:02 crc kubenswrapper[4967]: I1008 00:08:02.764267 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:02Z","lastTransitionTime":"2025-10-08T00:08:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:02 crc kubenswrapper[4967]: I1008 00:08:02.867030 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:02 crc kubenswrapper[4967]: I1008 00:08:02.867078 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:02 crc kubenswrapper[4967]: I1008 00:08:02.867091 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:02 crc kubenswrapper[4967]: I1008 00:08:02.867107 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:02 crc kubenswrapper[4967]: I1008 00:08:02.867118 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:02Z","lastTransitionTime":"2025-10-08T00:08:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:02 crc kubenswrapper[4967]: I1008 00:08:02.970164 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:02 crc kubenswrapper[4967]: I1008 00:08:02.970211 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:02 crc kubenswrapper[4967]: I1008 00:08:02.970260 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:02 crc kubenswrapper[4967]: I1008 00:08:02.970283 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:02 crc kubenswrapper[4967]: I1008 00:08:02.970303 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:02Z","lastTransitionTime":"2025-10-08T00:08:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:03 crc kubenswrapper[4967]: I1008 00:08:03.072835 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:03 crc kubenswrapper[4967]: I1008 00:08:03.072934 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:03 crc kubenswrapper[4967]: I1008 00:08:03.072966 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:03 crc kubenswrapper[4967]: I1008 00:08:03.073001 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:03 crc kubenswrapper[4967]: I1008 00:08:03.073025 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:03Z","lastTransitionTime":"2025-10-08T00:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:03 crc kubenswrapper[4967]: I1008 00:08:03.117476 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 00:08:03 crc kubenswrapper[4967]: I1008 00:08:03.117542 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 00:08:03 crc kubenswrapper[4967]: E1008 00:08:03.117753 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 00:08:03 crc kubenswrapper[4967]: E1008 00:08:03.117981 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 00:08:03 crc kubenswrapper[4967]: I1008 00:08:03.176264 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:03 crc kubenswrapper[4967]: I1008 00:08:03.176352 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:03 crc kubenswrapper[4967]: I1008 00:08:03.176377 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:03 crc kubenswrapper[4967]: I1008 00:08:03.176403 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:03 crc kubenswrapper[4967]: I1008 00:08:03.176424 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:03Z","lastTransitionTime":"2025-10-08T00:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:03 crc kubenswrapper[4967]: I1008 00:08:03.279324 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:03 crc kubenswrapper[4967]: I1008 00:08:03.279410 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:03 crc kubenswrapper[4967]: I1008 00:08:03.279428 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:03 crc kubenswrapper[4967]: I1008 00:08:03.279459 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:03 crc kubenswrapper[4967]: I1008 00:08:03.279496 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:03Z","lastTransitionTime":"2025-10-08T00:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:03 crc kubenswrapper[4967]: I1008 00:08:03.381611 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:03 crc kubenswrapper[4967]: I1008 00:08:03.381726 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:03 crc kubenswrapper[4967]: I1008 00:08:03.381746 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:03 crc kubenswrapper[4967]: I1008 00:08:03.381772 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:03 crc kubenswrapper[4967]: I1008 00:08:03.381791 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:03Z","lastTransitionTime":"2025-10-08T00:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:03 crc kubenswrapper[4967]: I1008 00:08:03.484184 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:03 crc kubenswrapper[4967]: I1008 00:08:03.484247 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:03 crc kubenswrapper[4967]: I1008 00:08:03.484264 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:03 crc kubenswrapper[4967]: I1008 00:08:03.484288 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:03 crc kubenswrapper[4967]: I1008 00:08:03.484307 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:03Z","lastTransitionTime":"2025-10-08T00:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:03 crc kubenswrapper[4967]: I1008 00:08:03.586878 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:03 crc kubenswrapper[4967]: I1008 00:08:03.586955 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:03 crc kubenswrapper[4967]: I1008 00:08:03.586976 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:03 crc kubenswrapper[4967]: I1008 00:08:03.587003 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:03 crc kubenswrapper[4967]: I1008 00:08:03.587024 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:03Z","lastTransitionTime":"2025-10-08T00:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:03 crc kubenswrapper[4967]: I1008 00:08:03.691010 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:03 crc kubenswrapper[4967]: I1008 00:08:03.691090 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:03 crc kubenswrapper[4967]: I1008 00:08:03.691112 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:03 crc kubenswrapper[4967]: I1008 00:08:03.691139 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:03 crc kubenswrapper[4967]: I1008 00:08:03.691160 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:03Z","lastTransitionTime":"2025-10-08T00:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:03 crc kubenswrapper[4967]: I1008 00:08:03.794159 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:03 crc kubenswrapper[4967]: I1008 00:08:03.794233 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:03 crc kubenswrapper[4967]: I1008 00:08:03.794271 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:03 crc kubenswrapper[4967]: I1008 00:08:03.794307 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:03 crc kubenswrapper[4967]: I1008 00:08:03.794330 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:03Z","lastTransitionTime":"2025-10-08T00:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:03 crc kubenswrapper[4967]: I1008 00:08:03.897297 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:03 crc kubenswrapper[4967]: I1008 00:08:03.897334 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:03 crc kubenswrapper[4967]: I1008 00:08:03.897346 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:03 crc kubenswrapper[4967]: I1008 00:08:03.897362 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:03 crc kubenswrapper[4967]: I1008 00:08:03.897374 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:03Z","lastTransitionTime":"2025-10-08T00:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:04 crc kubenswrapper[4967]: I1008 00:08:04.000207 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:04 crc kubenswrapper[4967]: I1008 00:08:04.000248 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:04 crc kubenswrapper[4967]: I1008 00:08:04.000259 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:04 crc kubenswrapper[4967]: I1008 00:08:04.000275 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:04 crc kubenswrapper[4967]: I1008 00:08:04.000286 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:04Z","lastTransitionTime":"2025-10-08T00:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:04 crc kubenswrapper[4967]: I1008 00:08:04.103027 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:04 crc kubenswrapper[4967]: I1008 00:08:04.103065 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:04 crc kubenswrapper[4967]: I1008 00:08:04.103075 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:04 crc kubenswrapper[4967]: I1008 00:08:04.103090 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:04 crc kubenswrapper[4967]: I1008 00:08:04.103099 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:04Z","lastTransitionTime":"2025-10-08T00:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:04 crc kubenswrapper[4967]: I1008 00:08:04.117824 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2h7z8" Oct 08 00:08:04 crc kubenswrapper[4967]: E1008 00:08:04.118014 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2h7z8" podUID="59f68771-e25d-4367-87ee-68a31d767882" Oct 08 00:08:04 crc kubenswrapper[4967]: I1008 00:08:04.118188 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 00:08:04 crc kubenswrapper[4967]: E1008 00:08:04.118377 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 00:08:04 crc kubenswrapper[4967]: I1008 00:08:04.207504 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:04 crc kubenswrapper[4967]: I1008 00:08:04.207559 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:04 crc kubenswrapper[4967]: I1008 00:08:04.207578 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:04 crc kubenswrapper[4967]: I1008 00:08:04.207607 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:04 crc kubenswrapper[4967]: I1008 00:08:04.207627 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:04Z","lastTransitionTime":"2025-10-08T00:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:04 crc kubenswrapper[4967]: I1008 00:08:04.310992 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:04 crc kubenswrapper[4967]: I1008 00:08:04.311043 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:04 crc kubenswrapper[4967]: I1008 00:08:04.311056 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:04 crc kubenswrapper[4967]: I1008 00:08:04.311075 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:04 crc kubenswrapper[4967]: I1008 00:08:04.311088 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:04Z","lastTransitionTime":"2025-10-08T00:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:04 crc kubenswrapper[4967]: I1008 00:08:04.413955 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:04 crc kubenswrapper[4967]: I1008 00:08:04.414325 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:04 crc kubenswrapper[4967]: I1008 00:08:04.414419 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:04 crc kubenswrapper[4967]: I1008 00:08:04.414515 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:04 crc kubenswrapper[4967]: I1008 00:08:04.414598 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:04Z","lastTransitionTime":"2025-10-08T00:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:04 crc kubenswrapper[4967]: I1008 00:08:04.517728 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:04 crc kubenswrapper[4967]: I1008 00:08:04.517771 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:04 crc kubenswrapper[4967]: I1008 00:08:04.517787 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:04 crc kubenswrapper[4967]: I1008 00:08:04.517805 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:04 crc kubenswrapper[4967]: I1008 00:08:04.517817 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:04Z","lastTransitionTime":"2025-10-08T00:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:04 crc kubenswrapper[4967]: I1008 00:08:04.620808 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:04 crc kubenswrapper[4967]: I1008 00:08:04.620894 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:04 crc kubenswrapper[4967]: I1008 00:08:04.620914 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:04 crc kubenswrapper[4967]: I1008 00:08:04.620944 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:04 crc kubenswrapper[4967]: I1008 00:08:04.620969 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:04Z","lastTransitionTime":"2025-10-08T00:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:04 crc kubenswrapper[4967]: I1008 00:08:04.724065 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:04 crc kubenswrapper[4967]: I1008 00:08:04.724117 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:04 crc kubenswrapper[4967]: I1008 00:08:04.724135 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:04 crc kubenswrapper[4967]: I1008 00:08:04.724160 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:04 crc kubenswrapper[4967]: I1008 00:08:04.724177 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:04Z","lastTransitionTime":"2025-10-08T00:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:04 crc kubenswrapper[4967]: I1008 00:08:04.827138 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:04 crc kubenswrapper[4967]: I1008 00:08:04.827179 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:04 crc kubenswrapper[4967]: I1008 00:08:04.827191 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:04 crc kubenswrapper[4967]: I1008 00:08:04.827207 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:04 crc kubenswrapper[4967]: I1008 00:08:04.827217 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:04Z","lastTransitionTime":"2025-10-08T00:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:04 crc kubenswrapper[4967]: I1008 00:08:04.930237 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:04 crc kubenswrapper[4967]: I1008 00:08:04.930285 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:04 crc kubenswrapper[4967]: I1008 00:08:04.930301 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:04 crc kubenswrapper[4967]: I1008 00:08:04.930318 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:04 crc kubenswrapper[4967]: I1008 00:08:04.930329 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:04Z","lastTransitionTime":"2025-10-08T00:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:05 crc kubenswrapper[4967]: I1008 00:08:05.033134 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:05 crc kubenswrapper[4967]: I1008 00:08:05.033394 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:05 crc kubenswrapper[4967]: I1008 00:08:05.033476 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:05 crc kubenswrapper[4967]: I1008 00:08:05.033643 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:05 crc kubenswrapper[4967]: I1008 00:08:05.033801 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:05Z","lastTransitionTime":"2025-10-08T00:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:05 crc kubenswrapper[4967]: I1008 00:08:05.117308 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 00:08:05 crc kubenswrapper[4967]: I1008 00:08:05.117474 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 00:08:05 crc kubenswrapper[4967]: E1008 00:08:05.117851 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 00:08:05 crc kubenswrapper[4967]: E1008 00:08:05.118019 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 00:08:05 crc kubenswrapper[4967]: I1008 00:08:05.136573 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:05 crc kubenswrapper[4967]: I1008 00:08:05.136895 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:05 crc kubenswrapper[4967]: I1008 00:08:05.137036 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:05 crc kubenswrapper[4967]: I1008 00:08:05.137161 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:05 crc kubenswrapper[4967]: I1008 00:08:05.137356 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:05Z","lastTransitionTime":"2025-10-08T00:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:05 crc kubenswrapper[4967]: I1008 00:08:05.239889 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:05 crc kubenswrapper[4967]: I1008 00:08:05.240325 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:05 crc kubenswrapper[4967]: I1008 00:08:05.240518 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:05 crc kubenswrapper[4967]: I1008 00:08:05.240738 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:05 crc kubenswrapper[4967]: I1008 00:08:05.240926 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:05Z","lastTransitionTime":"2025-10-08T00:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:05 crc kubenswrapper[4967]: I1008 00:08:05.343948 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:05 crc kubenswrapper[4967]: I1008 00:08:05.343995 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:05 crc kubenswrapper[4967]: I1008 00:08:05.344009 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:05 crc kubenswrapper[4967]: I1008 00:08:05.344034 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:05 crc kubenswrapper[4967]: I1008 00:08:05.344049 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:05Z","lastTransitionTime":"2025-10-08T00:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:05 crc kubenswrapper[4967]: I1008 00:08:05.447288 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:05 crc kubenswrapper[4967]: I1008 00:08:05.447598 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:05 crc kubenswrapper[4967]: I1008 00:08:05.447932 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:05 crc kubenswrapper[4967]: I1008 00:08:05.448171 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:05 crc kubenswrapper[4967]: I1008 00:08:05.448375 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:05Z","lastTransitionTime":"2025-10-08T00:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:05 crc kubenswrapper[4967]: I1008 00:08:05.551554 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:05 crc kubenswrapper[4967]: I1008 00:08:05.552027 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:05 crc kubenswrapper[4967]: I1008 00:08:05.552199 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:05 crc kubenswrapper[4967]: I1008 00:08:05.552340 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:05 crc kubenswrapper[4967]: I1008 00:08:05.552481 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:05Z","lastTransitionTime":"2025-10-08T00:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:05 crc kubenswrapper[4967]: I1008 00:08:05.660216 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:05 crc kubenswrapper[4967]: I1008 00:08:05.660297 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:05 crc kubenswrapper[4967]: I1008 00:08:05.660321 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:05 crc kubenswrapper[4967]: I1008 00:08:05.660352 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:05 crc kubenswrapper[4967]: I1008 00:08:05.660374 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:05Z","lastTransitionTime":"2025-10-08T00:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:05 crc kubenswrapper[4967]: I1008 00:08:05.764351 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:05 crc kubenswrapper[4967]: I1008 00:08:05.765355 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:05 crc kubenswrapper[4967]: I1008 00:08:05.765484 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:05 crc kubenswrapper[4967]: I1008 00:08:05.765722 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:05 crc kubenswrapper[4967]: I1008 00:08:05.765789 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:05Z","lastTransitionTime":"2025-10-08T00:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:05 crc kubenswrapper[4967]: I1008 00:08:05.869750 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:05 crc kubenswrapper[4967]: I1008 00:08:05.870186 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:05 crc kubenswrapper[4967]: I1008 00:08:05.870274 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:05 crc kubenswrapper[4967]: I1008 00:08:05.870385 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:05 crc kubenswrapper[4967]: I1008 00:08:05.870487 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:05Z","lastTransitionTime":"2025-10-08T00:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:05 crc kubenswrapper[4967]: I1008 00:08:05.974375 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:05 crc kubenswrapper[4967]: I1008 00:08:05.974441 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:05 crc kubenswrapper[4967]: I1008 00:08:05.974457 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:05 crc kubenswrapper[4967]: I1008 00:08:05.974480 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:05 crc kubenswrapper[4967]: I1008 00:08:05.974496 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:05Z","lastTransitionTime":"2025-10-08T00:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:06 crc kubenswrapper[4967]: I1008 00:08:06.077522 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:06 crc kubenswrapper[4967]: I1008 00:08:06.077589 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:06 crc kubenswrapper[4967]: I1008 00:08:06.077607 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:06 crc kubenswrapper[4967]: I1008 00:08:06.077632 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:06 crc kubenswrapper[4967]: I1008 00:08:06.077650 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:06Z","lastTransitionTime":"2025-10-08T00:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:06 crc kubenswrapper[4967]: I1008 00:08:06.117442 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2h7z8" Oct 08 00:08:06 crc kubenswrapper[4967]: I1008 00:08:06.117488 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 00:08:06 crc kubenswrapper[4967]: E1008 00:08:06.117644 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2h7z8" podUID="59f68771-e25d-4367-87ee-68a31d767882" Oct 08 00:08:06 crc kubenswrapper[4967]: E1008 00:08:06.117816 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 00:08:06 crc kubenswrapper[4967]: I1008 00:08:06.180649 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:06 crc kubenswrapper[4967]: I1008 00:08:06.180716 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:06 crc kubenswrapper[4967]: I1008 00:08:06.180731 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:06 crc kubenswrapper[4967]: I1008 00:08:06.180750 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:06 crc kubenswrapper[4967]: I1008 00:08:06.180762 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:06Z","lastTransitionTime":"2025-10-08T00:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:06 crc kubenswrapper[4967]: I1008 00:08:06.284269 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:06 crc kubenswrapper[4967]: I1008 00:08:06.284336 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:06 crc kubenswrapper[4967]: I1008 00:08:06.284357 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:06 crc kubenswrapper[4967]: I1008 00:08:06.284382 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:06 crc kubenswrapper[4967]: I1008 00:08:06.284401 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:06Z","lastTransitionTime":"2025-10-08T00:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:06 crc kubenswrapper[4967]: I1008 00:08:06.387789 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:06 crc kubenswrapper[4967]: I1008 00:08:06.387846 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:06 crc kubenswrapper[4967]: I1008 00:08:06.387863 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:06 crc kubenswrapper[4967]: I1008 00:08:06.387890 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:06 crc kubenswrapper[4967]: I1008 00:08:06.387908 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:06Z","lastTransitionTime":"2025-10-08T00:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:06 crc kubenswrapper[4967]: I1008 00:08:06.491644 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:06 crc kubenswrapper[4967]: I1008 00:08:06.491737 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:06 crc kubenswrapper[4967]: I1008 00:08:06.491757 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:06 crc kubenswrapper[4967]: I1008 00:08:06.491779 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:06 crc kubenswrapper[4967]: I1008 00:08:06.491797 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:06Z","lastTransitionTime":"2025-10-08T00:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:06 crc kubenswrapper[4967]: I1008 00:08:06.595113 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:06 crc kubenswrapper[4967]: I1008 00:08:06.595161 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:06 crc kubenswrapper[4967]: I1008 00:08:06.595173 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:06 crc kubenswrapper[4967]: I1008 00:08:06.595190 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:06 crc kubenswrapper[4967]: I1008 00:08:06.595203 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:06Z","lastTransitionTime":"2025-10-08T00:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:06 crc kubenswrapper[4967]: I1008 00:08:06.697573 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:06 crc kubenswrapper[4967]: I1008 00:08:06.697660 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:06 crc kubenswrapper[4967]: I1008 00:08:06.697720 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:06 crc kubenswrapper[4967]: I1008 00:08:06.697752 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:06 crc kubenswrapper[4967]: I1008 00:08:06.697776 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:06Z","lastTransitionTime":"2025-10-08T00:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:06 crc kubenswrapper[4967]: I1008 00:08:06.800463 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:06 crc kubenswrapper[4967]: I1008 00:08:06.800515 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:06 crc kubenswrapper[4967]: I1008 00:08:06.800526 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:06 crc kubenswrapper[4967]: I1008 00:08:06.800543 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:06 crc kubenswrapper[4967]: I1008 00:08:06.800554 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:06Z","lastTransitionTime":"2025-10-08T00:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:06 crc kubenswrapper[4967]: I1008 00:08:06.903304 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:06 crc kubenswrapper[4967]: I1008 00:08:06.903355 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:06 crc kubenswrapper[4967]: I1008 00:08:06.903369 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:06 crc kubenswrapper[4967]: I1008 00:08:06.903388 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:06 crc kubenswrapper[4967]: I1008 00:08:06.903399 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:06Z","lastTransitionTime":"2025-10-08T00:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:07 crc kubenswrapper[4967]: I1008 00:08:07.005629 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:07 crc kubenswrapper[4967]: I1008 00:08:07.005952 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:07 crc kubenswrapper[4967]: I1008 00:08:07.005962 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:07 crc kubenswrapper[4967]: I1008 00:08:07.005976 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:07 crc kubenswrapper[4967]: I1008 00:08:07.005988 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:07Z","lastTransitionTime":"2025-10-08T00:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:07 crc kubenswrapper[4967]: I1008 00:08:07.108516 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 00:08:07 crc kubenswrapper[4967]: I1008 00:08:07.109638 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:07 crc kubenswrapper[4967]: I1008 00:08:07.109701 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:07 crc kubenswrapper[4967]: I1008 00:08:07.109722 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:07 crc kubenswrapper[4967]: I1008 00:08:07.109738 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:07 crc kubenswrapper[4967]: I1008 00:08:07.109752 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:07Z","lastTransitionTime":"2025-10-08T00:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:07 crc kubenswrapper[4967]: E1008 00:08:07.108719 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 00:09:11.108652613 +0000 UTC m=+149.759187312 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:07 crc kubenswrapper[4967]: I1008 00:08:07.110294 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 00:08:07 crc kubenswrapper[4967]: I1008 00:08:07.110709 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 00:08:07 crc kubenswrapper[4967]: E1008 00:08:07.110764 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 00:08:07 crc kubenswrapper[4967]: E1008 00:08:07.110950 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 00:08:07 crc kubenswrapper[4967]: I1008 00:08:07.110897 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 00:08:07 crc kubenswrapper[4967]: E1008 00:08:07.110969 4967 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 00:08:07 crc kubenswrapper[4967]: I1008 00:08:07.111057 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 00:08:07 crc kubenswrapper[4967]: E1008 00:08:07.110811 4967 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 00:08:07 crc kubenswrapper[4967]: E1008 00:08:07.111121 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-08 00:09:11.111102263 +0000 UTC m=+149.761636922 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 00:08:07 crc kubenswrapper[4967]: E1008 00:08:07.111167 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 00:09:11.111146225 +0000 UTC m=+149.761680914 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 00:08:07 crc kubenswrapper[4967]: E1008 00:08:07.111230 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 00:08:07 crc kubenswrapper[4967]: E1008 00:08:07.111263 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 00:08:07 crc kubenswrapper[4967]: E1008 00:08:07.111283 4967 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 00:08:07 crc kubenswrapper[4967]: E1008 00:08:07.111350 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-08 00:09:11.111332439 +0000 UTC m=+149.761867128 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 00:08:07 crc kubenswrapper[4967]: E1008 00:08:07.111835 4967 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 00:08:07 crc kubenswrapper[4967]: E1008 00:08:07.111910 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 00:09:11.111893923 +0000 UTC m=+149.762428662 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 00:08:07 crc kubenswrapper[4967]: I1008 00:08:07.117227 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 00:08:07 crc kubenswrapper[4967]: E1008 00:08:07.117409 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 00:08:07 crc kubenswrapper[4967]: I1008 00:08:07.117426 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 00:08:07 crc kubenswrapper[4967]: E1008 00:08:07.117963 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 00:08:07 crc kubenswrapper[4967]: I1008 00:08:07.215964 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:07 crc kubenswrapper[4967]: I1008 00:08:07.216002 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:07 crc kubenswrapper[4967]: I1008 00:08:07.216025 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:07 crc kubenswrapper[4967]: I1008 00:08:07.216047 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:07 crc kubenswrapper[4967]: I1008 00:08:07.216061 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:07Z","lastTransitionTime":"2025-10-08T00:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:07 crc kubenswrapper[4967]: I1008 00:08:07.319475 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:07 crc kubenswrapper[4967]: I1008 00:08:07.319533 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:07 crc kubenswrapper[4967]: I1008 00:08:07.319551 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:07 crc kubenswrapper[4967]: I1008 00:08:07.319610 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:07 crc kubenswrapper[4967]: I1008 00:08:07.319631 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:07Z","lastTransitionTime":"2025-10-08T00:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:07 crc kubenswrapper[4967]: I1008 00:08:07.422803 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:07 crc kubenswrapper[4967]: I1008 00:08:07.422859 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:07 crc kubenswrapper[4967]: I1008 00:08:07.422875 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:07 crc kubenswrapper[4967]: I1008 00:08:07.422896 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:07 crc kubenswrapper[4967]: I1008 00:08:07.422912 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:07Z","lastTransitionTime":"2025-10-08T00:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:07 crc kubenswrapper[4967]: I1008 00:08:07.525750 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:07 crc kubenswrapper[4967]: I1008 00:08:07.525820 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:07 crc kubenswrapper[4967]: I1008 00:08:07.525847 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:07 crc kubenswrapper[4967]: I1008 00:08:07.525875 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:07 crc kubenswrapper[4967]: I1008 00:08:07.525897 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:07Z","lastTransitionTime":"2025-10-08T00:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:07 crc kubenswrapper[4967]: I1008 00:08:07.628248 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:07 crc kubenswrapper[4967]: I1008 00:08:07.628543 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:07 crc kubenswrapper[4967]: I1008 00:08:07.628638 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:07 crc kubenswrapper[4967]: I1008 00:08:07.628744 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:07 crc kubenswrapper[4967]: I1008 00:08:07.628834 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:07Z","lastTransitionTime":"2025-10-08T00:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:07 crc kubenswrapper[4967]: I1008 00:08:07.730798 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:07 crc kubenswrapper[4967]: I1008 00:08:07.731091 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:07 crc kubenswrapper[4967]: I1008 00:08:07.731237 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:07 crc kubenswrapper[4967]: I1008 00:08:07.731333 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:07 crc kubenswrapper[4967]: I1008 00:08:07.731430 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:07Z","lastTransitionTime":"2025-10-08T00:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:07 crc kubenswrapper[4967]: I1008 00:08:07.834998 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:07 crc kubenswrapper[4967]: I1008 00:08:07.835051 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:07 crc kubenswrapper[4967]: I1008 00:08:07.835067 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:07 crc kubenswrapper[4967]: I1008 00:08:07.835089 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:07 crc kubenswrapper[4967]: I1008 00:08:07.835103 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:07Z","lastTransitionTime":"2025-10-08T00:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:07 crc kubenswrapper[4967]: I1008 00:08:07.937278 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:07 crc kubenswrapper[4967]: I1008 00:08:07.937317 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:07 crc kubenswrapper[4967]: I1008 00:08:07.937330 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:07 crc kubenswrapper[4967]: I1008 00:08:07.937344 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:07 crc kubenswrapper[4967]: I1008 00:08:07.937353 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:07Z","lastTransitionTime":"2025-10-08T00:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:08 crc kubenswrapper[4967]: I1008 00:08:08.039583 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:08 crc kubenswrapper[4967]: I1008 00:08:08.039648 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:08 crc kubenswrapper[4967]: I1008 00:08:08.039749 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:08 crc kubenswrapper[4967]: I1008 00:08:08.039786 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:08 crc kubenswrapper[4967]: I1008 00:08:08.039844 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:08Z","lastTransitionTime":"2025-10-08T00:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:08 crc kubenswrapper[4967]: I1008 00:08:08.117767 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2h7z8" Oct 08 00:08:08 crc kubenswrapper[4967]: E1008 00:08:08.117901 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2h7z8" podUID="59f68771-e25d-4367-87ee-68a31d767882" Oct 08 00:08:08 crc kubenswrapper[4967]: I1008 00:08:08.118015 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 00:08:08 crc kubenswrapper[4967]: E1008 00:08:08.118180 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 00:08:08 crc kubenswrapper[4967]: I1008 00:08:08.142780 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:08 crc kubenswrapper[4967]: I1008 00:08:08.142844 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:08 crc kubenswrapper[4967]: I1008 00:08:08.142864 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:08 crc kubenswrapper[4967]: I1008 00:08:08.142885 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:08 crc kubenswrapper[4967]: I1008 00:08:08.142902 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:08Z","lastTransitionTime":"2025-10-08T00:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:08 crc kubenswrapper[4967]: I1008 00:08:08.245819 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:08 crc kubenswrapper[4967]: I1008 00:08:08.245866 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:08 crc kubenswrapper[4967]: I1008 00:08:08.245880 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:08 crc kubenswrapper[4967]: I1008 00:08:08.246333 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:08 crc kubenswrapper[4967]: I1008 00:08:08.246416 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:08Z","lastTransitionTime":"2025-10-08T00:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:08 crc kubenswrapper[4967]: I1008 00:08:08.349637 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:08 crc kubenswrapper[4967]: I1008 00:08:08.349751 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:08 crc kubenswrapper[4967]: I1008 00:08:08.349776 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:08 crc kubenswrapper[4967]: I1008 00:08:08.349817 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:08 crc kubenswrapper[4967]: I1008 00:08:08.349836 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:08Z","lastTransitionTime":"2025-10-08T00:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:08 crc kubenswrapper[4967]: I1008 00:08:08.452350 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:08 crc kubenswrapper[4967]: I1008 00:08:08.452390 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:08 crc kubenswrapper[4967]: I1008 00:08:08.452403 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:08 crc kubenswrapper[4967]: I1008 00:08:08.452421 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:08 crc kubenswrapper[4967]: I1008 00:08:08.452433 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:08Z","lastTransitionTime":"2025-10-08T00:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:08 crc kubenswrapper[4967]: I1008 00:08:08.555379 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:08 crc kubenswrapper[4967]: I1008 00:08:08.555764 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:08 crc kubenswrapper[4967]: I1008 00:08:08.555898 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:08 crc kubenswrapper[4967]: I1008 00:08:08.556008 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:08 crc kubenswrapper[4967]: I1008 00:08:08.556076 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:08Z","lastTransitionTime":"2025-10-08T00:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:08 crc kubenswrapper[4967]: I1008 00:08:08.659145 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:08 crc kubenswrapper[4967]: I1008 00:08:08.659213 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:08 crc kubenswrapper[4967]: I1008 00:08:08.659231 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:08 crc kubenswrapper[4967]: I1008 00:08:08.659258 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:08 crc kubenswrapper[4967]: I1008 00:08:08.659275 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:08Z","lastTransitionTime":"2025-10-08T00:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:08 crc kubenswrapper[4967]: I1008 00:08:08.762567 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:08 crc kubenswrapper[4967]: I1008 00:08:08.762617 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:08 crc kubenswrapper[4967]: I1008 00:08:08.762632 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:08 crc kubenswrapper[4967]: I1008 00:08:08.762650 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:08 crc kubenswrapper[4967]: I1008 00:08:08.762662 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:08Z","lastTransitionTime":"2025-10-08T00:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:08 crc kubenswrapper[4967]: I1008 00:08:08.865865 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:08 crc kubenswrapper[4967]: I1008 00:08:08.865950 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:08 crc kubenswrapper[4967]: I1008 00:08:08.865974 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:08 crc kubenswrapper[4967]: I1008 00:08:08.866009 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:08 crc kubenswrapper[4967]: I1008 00:08:08.866029 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:08Z","lastTransitionTime":"2025-10-08T00:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:08 crc kubenswrapper[4967]: I1008 00:08:08.969904 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:08 crc kubenswrapper[4967]: I1008 00:08:08.969960 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:08 crc kubenswrapper[4967]: I1008 00:08:08.969974 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:08 crc kubenswrapper[4967]: I1008 00:08:08.969996 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:08 crc kubenswrapper[4967]: I1008 00:08:08.970011 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:08Z","lastTransitionTime":"2025-10-08T00:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:09 crc kubenswrapper[4967]: I1008 00:08:09.073433 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:09 crc kubenswrapper[4967]: I1008 00:08:09.073965 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:09 crc kubenswrapper[4967]: I1008 00:08:09.074125 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:09 crc kubenswrapper[4967]: I1008 00:08:09.074267 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:09 crc kubenswrapper[4967]: I1008 00:08:09.074470 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:09Z","lastTransitionTime":"2025-10-08T00:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:09 crc kubenswrapper[4967]: I1008 00:08:09.117335 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 00:08:09 crc kubenswrapper[4967]: I1008 00:08:09.117345 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 00:08:09 crc kubenswrapper[4967]: E1008 00:08:09.117804 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 00:08:09 crc kubenswrapper[4967]: E1008 00:08:09.117935 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 00:08:09 crc kubenswrapper[4967]: I1008 00:08:09.178146 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:09 crc kubenswrapper[4967]: I1008 00:08:09.178223 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:09 crc kubenswrapper[4967]: I1008 00:08:09.178246 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:09 crc kubenswrapper[4967]: I1008 00:08:09.178281 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:09 crc kubenswrapper[4967]: I1008 00:08:09.178306 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:09Z","lastTransitionTime":"2025-10-08T00:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:09 crc kubenswrapper[4967]: I1008 00:08:09.281748 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:09 crc kubenswrapper[4967]: I1008 00:08:09.281823 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:09 crc kubenswrapper[4967]: I1008 00:08:09.281838 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:09 crc kubenswrapper[4967]: I1008 00:08:09.281868 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:09 crc kubenswrapper[4967]: I1008 00:08:09.281889 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:09Z","lastTransitionTime":"2025-10-08T00:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:09 crc kubenswrapper[4967]: I1008 00:08:09.385751 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:09 crc kubenswrapper[4967]: I1008 00:08:09.385799 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:09 crc kubenswrapper[4967]: I1008 00:08:09.385809 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:09 crc kubenswrapper[4967]: I1008 00:08:09.385826 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:09 crc kubenswrapper[4967]: I1008 00:08:09.385838 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:09Z","lastTransitionTime":"2025-10-08T00:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:09 crc kubenswrapper[4967]: I1008 00:08:09.493475 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:09 crc kubenswrapper[4967]: I1008 00:08:09.493570 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:09 crc kubenswrapper[4967]: I1008 00:08:09.493586 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:09 crc kubenswrapper[4967]: I1008 00:08:09.493615 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:09 crc kubenswrapper[4967]: I1008 00:08:09.493631 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:09Z","lastTransitionTime":"2025-10-08T00:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:09 crc kubenswrapper[4967]: I1008 00:08:09.596393 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:09 crc kubenswrapper[4967]: I1008 00:08:09.596441 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:09 crc kubenswrapper[4967]: I1008 00:08:09.596453 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:09 crc kubenswrapper[4967]: I1008 00:08:09.596472 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:09 crc kubenswrapper[4967]: I1008 00:08:09.596484 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:09Z","lastTransitionTime":"2025-10-08T00:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:09 crc kubenswrapper[4967]: I1008 00:08:09.698915 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:09 crc kubenswrapper[4967]: I1008 00:08:09.698977 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:09 crc kubenswrapper[4967]: I1008 00:08:09.698995 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:09 crc kubenswrapper[4967]: I1008 00:08:09.699020 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:09 crc kubenswrapper[4967]: I1008 00:08:09.699037 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:09Z","lastTransitionTime":"2025-10-08T00:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:09 crc kubenswrapper[4967]: I1008 00:08:09.802009 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:09 crc kubenswrapper[4967]: I1008 00:08:09.802084 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:09 crc kubenswrapper[4967]: I1008 00:08:09.802108 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:09 crc kubenswrapper[4967]: I1008 00:08:09.802139 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:09 crc kubenswrapper[4967]: I1008 00:08:09.802163 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:09Z","lastTransitionTime":"2025-10-08T00:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:09 crc kubenswrapper[4967]: I1008 00:08:09.904910 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:09 crc kubenswrapper[4967]: I1008 00:08:09.904956 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:09 crc kubenswrapper[4967]: I1008 00:08:09.904974 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:09 crc kubenswrapper[4967]: I1008 00:08:09.904996 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:09 crc kubenswrapper[4967]: I1008 00:08:09.905012 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:09Z","lastTransitionTime":"2025-10-08T00:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:10 crc kubenswrapper[4967]: I1008 00:08:10.007869 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:10 crc kubenswrapper[4967]: I1008 00:08:10.007915 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:10 crc kubenswrapper[4967]: I1008 00:08:10.007937 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:10 crc kubenswrapper[4967]: I1008 00:08:10.007967 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:10 crc kubenswrapper[4967]: I1008 00:08:10.007989 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:10Z","lastTransitionTime":"2025-10-08T00:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:10 crc kubenswrapper[4967]: I1008 00:08:10.110987 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:10 crc kubenswrapper[4967]: I1008 00:08:10.111052 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:10 crc kubenswrapper[4967]: I1008 00:08:10.111070 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:10 crc kubenswrapper[4967]: I1008 00:08:10.111096 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:10 crc kubenswrapper[4967]: I1008 00:08:10.111115 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:10Z","lastTransitionTime":"2025-10-08T00:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:10 crc kubenswrapper[4967]: I1008 00:08:10.117399 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2h7z8" Oct 08 00:08:10 crc kubenswrapper[4967]: I1008 00:08:10.117412 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 00:08:10 crc kubenswrapper[4967]: E1008 00:08:10.117576 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2h7z8" podUID="59f68771-e25d-4367-87ee-68a31d767882" Oct 08 00:08:10 crc kubenswrapper[4967]: E1008 00:08:10.117758 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 00:08:10 crc kubenswrapper[4967]: I1008 00:08:10.156992 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 00:08:10 crc kubenswrapper[4967]: I1008 00:08:10.157412 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 00:08:10 crc kubenswrapper[4967]: I1008 00:08:10.157650 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 00:08:10 crc kubenswrapper[4967]: I1008 00:08:10.157913 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 00:08:10 crc kubenswrapper[4967]: I1008 00:08:10.158061 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T00:08:10Z","lastTransitionTime":"2025-10-08T00:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 00:08:10 crc kubenswrapper[4967]: I1008 00:08:10.223727 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-mcn2x"] Oct 08 00:08:10 crc kubenswrapper[4967]: I1008 00:08:10.224524 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mcn2x" Oct 08 00:08:10 crc kubenswrapper[4967]: I1008 00:08:10.226880 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Oct 08 00:08:10 crc kubenswrapper[4967]: I1008 00:08:10.227240 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Oct 08 00:08:10 crc kubenswrapper[4967]: I1008 00:08:10.228100 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Oct 08 00:08:10 crc kubenswrapper[4967]: I1008 00:08:10.228190 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Oct 08 00:08:10 crc kubenswrapper[4967]: I1008 00:08:10.262532 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=36.262516249 podStartE2EDuration="36.262516249s" podCreationTimestamp="2025-10-08 00:07:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:08:10.26139093 +0000 UTC m=+88.911925579" watchObservedRunningTime="2025-10-08 00:08:10.262516249 +0000 UTC m=+88.913050898" Oct 08 00:08:10 crc kubenswrapper[4967]: I1008 00:08:10.291533 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-npz2s" podStartSLOduration=67.291513478 podStartE2EDuration="1m7.291513478s" podCreationTimestamp="2025-10-08 00:07:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:08:10.291077457 +0000 UTC m=+88.941612126" watchObservedRunningTime="2025-10-08 00:08:10.291513478 +0000 UTC m=+88.942048127" Oct 08 00:08:10 crc kubenswrapper[4967]: I1008 00:08:10.320744 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kpmlg" podStartSLOduration=66.320721753 podStartE2EDuration="1m6.320721753s" podCreationTimestamp="2025-10-08 00:07:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:08:10.32024677 +0000 UTC m=+88.970781489" watchObservedRunningTime="2025-10-08 00:08:10.320721753 +0000 UTC m=+88.971256412" Oct 08 00:08:10 crc kubenswrapper[4967]: I1008 00:08:10.348417 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/1b174bc2-3b3a-422a-92fc-843f26176149-service-ca\") pod \"cluster-version-operator-5c965bbfc6-mcn2x\" (UID: \"1b174bc2-3b3a-422a-92fc-843f26176149\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mcn2x" Oct 08 00:08:10 crc kubenswrapper[4967]: I1008 00:08:10.348745 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1b174bc2-3b3a-422a-92fc-843f26176149-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-mcn2x\" (UID: \"1b174bc2-3b3a-422a-92fc-843f26176149\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mcn2x" Oct 08 00:08:10 crc kubenswrapper[4967]: I1008 00:08:10.348870 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1b174bc2-3b3a-422a-92fc-843f26176149-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-mcn2x\" (UID: \"1b174bc2-3b3a-422a-92fc-843f26176149\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mcn2x" Oct 08 00:08:10 crc kubenswrapper[4967]: I1008 00:08:10.348974 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/1b174bc2-3b3a-422a-92fc-843f26176149-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-mcn2x\" (UID: \"1b174bc2-3b3a-422a-92fc-843f26176149\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mcn2x" Oct 08 00:08:10 crc kubenswrapper[4967]: I1008 00:08:10.349094 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/1b174bc2-3b3a-422a-92fc-843f26176149-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-mcn2x\" (UID: \"1b174bc2-3b3a-422a-92fc-843f26176149\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mcn2x" Oct 08 00:08:10 crc kubenswrapper[4967]: I1008 00:08:10.351389 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=68.351368723 podStartE2EDuration="1m8.351368723s" podCreationTimestamp="2025-10-08 00:07:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:08:10.338066303 +0000 UTC m=+88.988600962" watchObservedRunningTime="2025-10-08 00:08:10.351368723 +0000 UTC m=+89.001903372" Oct 08 00:08:10 crc kubenswrapper[4967]: I1008 00:08:10.390117 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-9mrr7" podStartSLOduration=67.390100223 podStartE2EDuration="1m7.390100223s" podCreationTimestamp="2025-10-08 00:07:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:08:10.376793623 +0000 UTC m=+89.027328282" watchObservedRunningTime="2025-10-08 00:08:10.390100223 +0000 UTC m=+89.040634872" Oct 08 00:08:10 crc kubenswrapper[4967]: I1008 00:08:10.390618 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" podStartSLOduration=67.390612786 podStartE2EDuration="1m7.390612786s" podCreationTimestamp="2025-10-08 00:07:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:08:10.389658412 +0000 UTC m=+89.040193071" watchObservedRunningTime="2025-10-08 00:08:10.390612786 +0000 UTC m=+89.041147435" Oct 08 00:08:10 crc kubenswrapper[4967]: I1008 00:08:10.415449 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=67.415434101 podStartE2EDuration="1m7.415434101s" podCreationTimestamp="2025-10-08 00:07:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:08:10.405830303 +0000 UTC m=+89.056365022" watchObservedRunningTime="2025-10-08 00:08:10.415434101 +0000 UTC m=+89.065968740" Oct 08 00:08:10 crc kubenswrapper[4967]: I1008 00:08:10.432581 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=23.432567897 podStartE2EDuration="23.432567897s" podCreationTimestamp="2025-10-08 00:07:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:08:10.416635851 +0000 UTC m=+89.067170510" watchObservedRunningTime="2025-10-08 00:08:10.432567897 +0000 UTC m=+89.083102546" Oct 08 00:08:10 crc kubenswrapper[4967]: I1008 00:08:10.449996 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1b174bc2-3b3a-422a-92fc-843f26176149-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-mcn2x\" (UID: \"1b174bc2-3b3a-422a-92fc-843f26176149\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mcn2x" Oct 08 00:08:10 crc kubenswrapper[4967]: I1008 00:08:10.450258 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/1b174bc2-3b3a-422a-92fc-843f26176149-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-mcn2x\" (UID: \"1b174bc2-3b3a-422a-92fc-843f26176149\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mcn2x" Oct 08 00:08:10 crc kubenswrapper[4967]: I1008 00:08:10.450352 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/1b174bc2-3b3a-422a-92fc-843f26176149-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-mcn2x\" (UID: \"1b174bc2-3b3a-422a-92fc-843f26176149\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mcn2x" Oct 08 00:08:10 crc kubenswrapper[4967]: I1008 00:08:10.450427 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/1b174bc2-3b3a-422a-92fc-843f26176149-service-ca\") pod \"cluster-version-operator-5c965bbfc6-mcn2x\" (UID: \"1b174bc2-3b3a-422a-92fc-843f26176149\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mcn2x" Oct 08 00:08:10 crc kubenswrapper[4967]: I1008 00:08:10.450371 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/1b174bc2-3b3a-422a-92fc-843f26176149-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-mcn2x\" (UID: \"1b174bc2-3b3a-422a-92fc-843f26176149\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mcn2x" Oct 08 00:08:10 crc kubenswrapper[4967]: I1008 00:08:10.450398 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/1b174bc2-3b3a-422a-92fc-843f26176149-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-mcn2x\" (UID: \"1b174bc2-3b3a-422a-92fc-843f26176149\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mcn2x" Oct 08 00:08:10 crc kubenswrapper[4967]: I1008 00:08:10.450504 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1b174bc2-3b3a-422a-92fc-843f26176149-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-mcn2x\" (UID: \"1b174bc2-3b3a-422a-92fc-843f26176149\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mcn2x" Oct 08 00:08:10 crc kubenswrapper[4967]: I1008 00:08:10.451157 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/1b174bc2-3b3a-422a-92fc-843f26176149-service-ca\") pod \"cluster-version-operator-5c965bbfc6-mcn2x\" (UID: \"1b174bc2-3b3a-422a-92fc-843f26176149\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mcn2x" Oct 08 00:08:10 crc kubenswrapper[4967]: I1008 00:08:10.456996 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1b174bc2-3b3a-422a-92fc-843f26176149-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-mcn2x\" (UID: \"1b174bc2-3b3a-422a-92fc-843f26176149\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mcn2x" Oct 08 00:08:10 crc kubenswrapper[4967]: I1008 00:08:10.457753 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-v5n7t" podStartSLOduration=67.457734081 podStartE2EDuration="1m7.457734081s" podCreationTimestamp="2025-10-08 00:07:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:08:10.457655149 +0000 UTC m=+89.108189808" watchObservedRunningTime="2025-10-08 00:08:10.457734081 +0000 UTC m=+89.108268780" Oct 08 00:08:10 crc kubenswrapper[4967]: I1008 00:08:10.480814 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1b174bc2-3b3a-422a-92fc-843f26176149-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-mcn2x\" (UID: \"1b174bc2-3b3a-422a-92fc-843f26176149\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mcn2x" Oct 08 00:08:10 crc kubenswrapper[4967]: I1008 00:08:10.501164 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-9tdbb" podStartSLOduration=67.501144527 podStartE2EDuration="1m7.501144527s" podCreationTimestamp="2025-10-08 00:07:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:08:10.500524482 +0000 UTC m=+89.151059141" watchObservedRunningTime="2025-10-08 00:08:10.501144527 +0000 UTC m=+89.151679186" Oct 08 00:08:10 crc kubenswrapper[4967]: I1008 00:08:10.544396 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mcn2x" Oct 08 00:08:10 crc kubenswrapper[4967]: I1008 00:08:10.704538 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mcn2x" event={"ID":"1b174bc2-3b3a-422a-92fc-843f26176149","Type":"ContainerStarted","Data":"828062b3ea12459db9a78ac629a4532484b0236ef93bae4b71e7df6c5de7825d"} Oct 08 00:08:10 crc kubenswrapper[4967]: I1008 00:08:10.705100 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mcn2x" event={"ID":"1b174bc2-3b3a-422a-92fc-843f26176149","Type":"ContainerStarted","Data":"67929f4e02cda918c8026043085953b838fd70f84be8ac3545754367d437f49e"} Oct 08 00:08:10 crc kubenswrapper[4967]: I1008 00:08:10.723500 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-mcn2x" podStartSLOduration=67.723474841 podStartE2EDuration="1m7.723474841s" podCreationTimestamp="2025-10-08 00:07:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:08:10.720006996 +0000 UTC m=+89.370541655" watchObservedRunningTime="2025-10-08 00:08:10.723474841 +0000 UTC m=+89.374009500" Oct 08 00:08:11 crc kubenswrapper[4967]: I1008 00:08:11.116918 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 00:08:11 crc kubenswrapper[4967]: I1008 00:08:11.117329 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 00:08:11 crc kubenswrapper[4967]: E1008 00:08:11.117548 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 00:08:11 crc kubenswrapper[4967]: I1008 00:08:11.117704 4967 scope.go:117] "RemoveContainer" containerID="4f7d46ee97fd3c4286f45879fededea788052cf83ebeebcb539be452f2ae102d" Oct 08 00:08:11 crc kubenswrapper[4967]: E1008 00:08:11.117869 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 00:08:11 crc kubenswrapper[4967]: E1008 00:08:11.117880 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-bs8w9_openshift-ovn-kubernetes(4f1af2ba-8907-4d14-afa1-ffde207c80f4)\"" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" podUID="4f1af2ba-8907-4d14-afa1-ffde207c80f4" Oct 08 00:08:12 crc kubenswrapper[4967]: I1008 00:08:12.117446 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 00:08:12 crc kubenswrapper[4967]: I1008 00:08:12.119653 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2h7z8" Oct 08 00:08:12 crc kubenswrapper[4967]: E1008 00:08:12.119845 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2h7z8" podUID="59f68771-e25d-4367-87ee-68a31d767882" Oct 08 00:08:12 crc kubenswrapper[4967]: E1008 00:08:12.120130 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 00:08:13 crc kubenswrapper[4967]: I1008 00:08:13.117483 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 00:08:13 crc kubenswrapper[4967]: E1008 00:08:13.117702 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 00:08:13 crc kubenswrapper[4967]: I1008 00:08:13.118295 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 00:08:13 crc kubenswrapper[4967]: E1008 00:08:13.118536 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 00:08:14 crc kubenswrapper[4967]: I1008 00:08:14.117546 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2h7z8" Oct 08 00:08:14 crc kubenswrapper[4967]: I1008 00:08:14.117576 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 00:08:14 crc kubenswrapper[4967]: E1008 00:08:14.118506 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2h7z8" podUID="59f68771-e25d-4367-87ee-68a31d767882" Oct 08 00:08:14 crc kubenswrapper[4967]: E1008 00:08:14.118706 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 00:08:15 crc kubenswrapper[4967]: I1008 00:08:15.117182 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 00:08:15 crc kubenswrapper[4967]: I1008 00:08:15.117193 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 00:08:15 crc kubenswrapper[4967]: E1008 00:08:15.117363 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 00:08:15 crc kubenswrapper[4967]: E1008 00:08:15.117418 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 00:08:16 crc kubenswrapper[4967]: I1008 00:08:16.117260 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 00:08:16 crc kubenswrapper[4967]: E1008 00:08:16.117415 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 00:08:16 crc kubenswrapper[4967]: I1008 00:08:16.117649 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2h7z8" Oct 08 00:08:16 crc kubenswrapper[4967]: E1008 00:08:16.117997 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2h7z8" podUID="59f68771-e25d-4367-87ee-68a31d767882" Oct 08 00:08:17 crc kubenswrapper[4967]: I1008 00:08:17.117853 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 00:08:17 crc kubenswrapper[4967]: I1008 00:08:17.117877 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 00:08:17 crc kubenswrapper[4967]: E1008 00:08:17.118016 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 00:08:17 crc kubenswrapper[4967]: E1008 00:08:17.118272 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 00:08:17 crc kubenswrapper[4967]: I1008 00:08:17.142849 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Oct 08 00:08:18 crc kubenswrapper[4967]: I1008 00:08:18.117790 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2h7z8" Oct 08 00:08:18 crc kubenswrapper[4967]: I1008 00:08:18.117802 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 00:08:18 crc kubenswrapper[4967]: E1008 00:08:18.118027 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2h7z8" podUID="59f68771-e25d-4367-87ee-68a31d767882" Oct 08 00:08:18 crc kubenswrapper[4967]: E1008 00:08:18.118086 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 00:08:19 crc kubenswrapper[4967]: I1008 00:08:19.117182 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 00:08:19 crc kubenswrapper[4967]: I1008 00:08:19.117250 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 00:08:19 crc kubenswrapper[4967]: E1008 00:08:19.117332 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 00:08:19 crc kubenswrapper[4967]: E1008 00:08:19.117392 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 00:08:20 crc kubenswrapper[4967]: I1008 00:08:20.117469 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2h7z8" Oct 08 00:08:20 crc kubenswrapper[4967]: I1008 00:08:20.117585 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 00:08:20 crc kubenswrapper[4967]: E1008 00:08:20.117770 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2h7z8" podUID="59f68771-e25d-4367-87ee-68a31d767882" Oct 08 00:08:20 crc kubenswrapper[4967]: E1008 00:08:20.117911 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 00:08:21 crc kubenswrapper[4967]: I1008 00:08:21.117400 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 00:08:21 crc kubenswrapper[4967]: E1008 00:08:21.117596 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 00:08:21 crc kubenswrapper[4967]: I1008 00:08:21.118506 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 00:08:21 crc kubenswrapper[4967]: E1008 00:08:21.118936 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 00:08:22 crc kubenswrapper[4967]: I1008 00:08:22.117704 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2h7z8" Oct 08 00:08:22 crc kubenswrapper[4967]: I1008 00:08:22.117740 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 00:08:22 crc kubenswrapper[4967]: E1008 00:08:22.119806 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2h7z8" podUID="59f68771-e25d-4367-87ee-68a31d767882" Oct 08 00:08:22 crc kubenswrapper[4967]: E1008 00:08:22.119918 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 00:08:22 crc kubenswrapper[4967]: I1008 00:08:22.160467 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=5.160444833 podStartE2EDuration="5.160444833s" podCreationTimestamp="2025-10-08 00:08:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:08:22.158421553 +0000 UTC m=+100.808956232" watchObservedRunningTime="2025-10-08 00:08:22.160444833 +0000 UTC m=+100.810979512" Oct 08 00:08:22 crc kubenswrapper[4967]: I1008 00:08:22.980225 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/59f68771-e25d-4367-87ee-68a31d767882-metrics-certs\") pod \"network-metrics-daemon-2h7z8\" (UID: \"59f68771-e25d-4367-87ee-68a31d767882\") " pod="openshift-multus/network-metrics-daemon-2h7z8" Oct 08 00:08:22 crc kubenswrapper[4967]: E1008 00:08:22.980584 4967 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 00:08:22 crc kubenswrapper[4967]: E1008 00:08:22.980731 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/59f68771-e25d-4367-87ee-68a31d767882-metrics-certs podName:59f68771-e25d-4367-87ee-68a31d767882 nodeName:}" failed. No retries permitted until 2025-10-08 00:09:26.980714929 +0000 UTC m=+165.631249578 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/59f68771-e25d-4367-87ee-68a31d767882-metrics-certs") pod "network-metrics-daemon-2h7z8" (UID: "59f68771-e25d-4367-87ee-68a31d767882") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 00:08:23 crc kubenswrapper[4967]: I1008 00:08:23.117295 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 00:08:23 crc kubenswrapper[4967]: I1008 00:08:23.117298 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 00:08:23 crc kubenswrapper[4967]: E1008 00:08:23.117443 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 00:08:23 crc kubenswrapper[4967]: E1008 00:08:23.117543 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 00:08:23 crc kubenswrapper[4967]: I1008 00:08:23.122714 4967 scope.go:117] "RemoveContainer" containerID="4f7d46ee97fd3c4286f45879fededea788052cf83ebeebcb539be452f2ae102d" Oct 08 00:08:23 crc kubenswrapper[4967]: E1008 00:08:23.123530 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-bs8w9_openshift-ovn-kubernetes(4f1af2ba-8907-4d14-afa1-ffde207c80f4)\"" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" podUID="4f1af2ba-8907-4d14-afa1-ffde207c80f4" Oct 08 00:08:24 crc kubenswrapper[4967]: I1008 00:08:24.117852 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2h7z8" Oct 08 00:08:24 crc kubenswrapper[4967]: E1008 00:08:24.118093 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2h7z8" podUID="59f68771-e25d-4367-87ee-68a31d767882" Oct 08 00:08:24 crc kubenswrapper[4967]: I1008 00:08:24.118164 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 00:08:24 crc kubenswrapper[4967]: E1008 00:08:24.118354 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 00:08:25 crc kubenswrapper[4967]: I1008 00:08:25.117838 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 00:08:25 crc kubenswrapper[4967]: I1008 00:08:25.117887 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 00:08:25 crc kubenswrapper[4967]: E1008 00:08:25.118146 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 00:08:25 crc kubenswrapper[4967]: E1008 00:08:25.118248 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 00:08:26 crc kubenswrapper[4967]: I1008 00:08:26.116916 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 00:08:26 crc kubenswrapper[4967]: I1008 00:08:26.116918 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2h7z8" Oct 08 00:08:26 crc kubenswrapper[4967]: E1008 00:08:26.117092 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 00:08:26 crc kubenswrapper[4967]: E1008 00:08:26.117507 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2h7z8" podUID="59f68771-e25d-4367-87ee-68a31d767882" Oct 08 00:08:27 crc kubenswrapper[4967]: I1008 00:08:27.117760 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 00:08:27 crc kubenswrapper[4967]: I1008 00:08:27.117800 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 00:08:27 crc kubenswrapper[4967]: E1008 00:08:27.117953 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 00:08:27 crc kubenswrapper[4967]: E1008 00:08:27.118146 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 00:08:28 crc kubenswrapper[4967]: I1008 00:08:28.116894 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2h7z8" Oct 08 00:08:28 crc kubenswrapper[4967]: I1008 00:08:28.116890 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 00:08:28 crc kubenswrapper[4967]: E1008 00:08:28.117036 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2h7z8" podUID="59f68771-e25d-4367-87ee-68a31d767882" Oct 08 00:08:28 crc kubenswrapper[4967]: E1008 00:08:28.117111 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 00:08:29 crc kubenswrapper[4967]: I1008 00:08:29.117033 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 00:08:29 crc kubenswrapper[4967]: I1008 00:08:29.117033 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 00:08:29 crc kubenswrapper[4967]: E1008 00:08:29.117345 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 00:08:29 crc kubenswrapper[4967]: E1008 00:08:29.117479 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 00:08:30 crc kubenswrapper[4967]: I1008 00:08:30.117502 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2h7z8" Oct 08 00:08:30 crc kubenswrapper[4967]: I1008 00:08:30.117501 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 00:08:30 crc kubenswrapper[4967]: E1008 00:08:30.117823 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 00:08:30 crc kubenswrapper[4967]: E1008 00:08:30.117704 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2h7z8" podUID="59f68771-e25d-4367-87ee-68a31d767882" Oct 08 00:08:31 crc kubenswrapper[4967]: I1008 00:08:31.117802 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 00:08:31 crc kubenswrapper[4967]: I1008 00:08:31.117802 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 00:08:31 crc kubenswrapper[4967]: E1008 00:08:31.118016 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 00:08:31 crc kubenswrapper[4967]: E1008 00:08:31.118175 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 00:08:32 crc kubenswrapper[4967]: I1008 00:08:32.116983 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2h7z8" Oct 08 00:08:32 crc kubenswrapper[4967]: I1008 00:08:32.117000 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 00:08:32 crc kubenswrapper[4967]: E1008 00:08:32.118725 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2h7z8" podUID="59f68771-e25d-4367-87ee-68a31d767882" Oct 08 00:08:32 crc kubenswrapper[4967]: E1008 00:08:32.118932 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 00:08:33 crc kubenswrapper[4967]: I1008 00:08:33.117304 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 00:08:33 crc kubenswrapper[4967]: I1008 00:08:33.117304 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 00:08:33 crc kubenswrapper[4967]: E1008 00:08:33.117797 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 00:08:33 crc kubenswrapper[4967]: E1008 00:08:33.117960 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 00:08:34 crc kubenswrapper[4967]: I1008 00:08:34.117745 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2h7z8" Oct 08 00:08:34 crc kubenswrapper[4967]: I1008 00:08:34.117804 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 00:08:34 crc kubenswrapper[4967]: E1008 00:08:34.118004 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2h7z8" podUID="59f68771-e25d-4367-87ee-68a31d767882" Oct 08 00:08:34 crc kubenswrapper[4967]: E1008 00:08:34.118107 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 00:08:35 crc kubenswrapper[4967]: I1008 00:08:35.117564 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 00:08:35 crc kubenswrapper[4967]: I1008 00:08:35.117619 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 00:08:35 crc kubenswrapper[4967]: E1008 00:08:35.117712 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 00:08:35 crc kubenswrapper[4967]: E1008 00:08:35.117857 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 00:08:36 crc kubenswrapper[4967]: I1008 00:08:36.117049 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2h7z8" Oct 08 00:08:36 crc kubenswrapper[4967]: I1008 00:08:36.117141 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 00:08:36 crc kubenswrapper[4967]: E1008 00:08:36.117288 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2h7z8" podUID="59f68771-e25d-4367-87ee-68a31d767882" Oct 08 00:08:36 crc kubenswrapper[4967]: E1008 00:08:36.117738 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 00:08:36 crc kubenswrapper[4967]: I1008 00:08:36.119266 4967 scope.go:117] "RemoveContainer" containerID="4f7d46ee97fd3c4286f45879fededea788052cf83ebeebcb539be452f2ae102d" Oct 08 00:08:36 crc kubenswrapper[4967]: I1008 00:08:36.795276 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bs8w9_4f1af2ba-8907-4d14-afa1-ffde207c80f4/ovnkube-controller/3.log" Oct 08 00:08:36 crc kubenswrapper[4967]: I1008 00:08:36.798237 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" event={"ID":"4f1af2ba-8907-4d14-afa1-ffde207c80f4","Type":"ContainerStarted","Data":"fab4c6588d3f2b44975eb2c248ad8b1cfef9b4b0fb797d0436123b63f9479df6"} Oct 08 00:08:36 crc kubenswrapper[4967]: I1008 00:08:36.798647 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" Oct 08 00:08:36 crc kubenswrapper[4967]: I1008 00:08:36.824166 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" podStartSLOduration=92.824147019 podStartE2EDuration="1m32.824147019s" podCreationTimestamp="2025-10-08 00:07:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:08:36.823553014 +0000 UTC m=+115.474087663" watchObservedRunningTime="2025-10-08 00:08:36.824147019 +0000 UTC m=+115.474681668" Oct 08 00:08:37 crc kubenswrapper[4967]: I1008 00:08:37.117057 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 00:08:37 crc kubenswrapper[4967]: E1008 00:08:37.117278 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 00:08:37 crc kubenswrapper[4967]: I1008 00:08:37.117403 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 00:08:37 crc kubenswrapper[4967]: E1008 00:08:37.117554 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 00:08:37 crc kubenswrapper[4967]: I1008 00:08:37.222208 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-2h7z8"] Oct 08 00:08:37 crc kubenswrapper[4967]: I1008 00:08:37.222324 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2h7z8" Oct 08 00:08:37 crc kubenswrapper[4967]: E1008 00:08:37.222426 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2h7z8" podUID="59f68771-e25d-4367-87ee-68a31d767882" Oct 08 00:08:38 crc kubenswrapper[4967]: I1008 00:08:38.117882 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 00:08:38 crc kubenswrapper[4967]: E1008 00:08:38.118472 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.117314 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.117393 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 00:08:39 crc kubenswrapper[4967]: E1008 00:08:39.117512 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.117536 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2h7z8" Oct 08 00:08:39 crc kubenswrapper[4967]: E1008 00:08:39.117608 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 00:08:39 crc kubenswrapper[4967]: E1008 00:08:39.117704 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-2h7z8" podUID="59f68771-e25d-4367-87ee-68a31d767882" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.547585 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.547840 4967 kubelet_node_status.go:538] "Fast updating node status as it just became ready" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.599476 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-vkf46"] Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.600001 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-vkf46" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.601162 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-v4kxw"] Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.601469 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v4kxw" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.605904 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-5wd5r"] Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.607724 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.607838 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.607974 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5wd5r" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.607735 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.608594 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.611794 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.612768 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.612793 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.613479 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.614419 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.614747 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.615959 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.617019 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-ck4pd"] Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.638372 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-ljjfc"] Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.638926 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-rjwhx"] Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.639540 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rjwhx" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.640444 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.645753 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-ljjfc" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.646291 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-ck4pd" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.646344 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-b5mzq"] Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.647349 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-b5mzq" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.648412 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.648485 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.648614 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.648744 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-p9jtz"] Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.648438 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.648985 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.649307 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-p9jtz" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.651602 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.652720 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.653080 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.653132 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.653455 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-tmj7c"] Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.653575 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.653935 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.653992 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.653882 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.654192 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-tmj7c" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.654047 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.654518 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.654801 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.655076 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.655168 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.655276 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.655296 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.660580 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.661314 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.661708 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.663152 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.663170 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.666753 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.667916 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.668076 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.668214 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.668446 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.668473 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.668815 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-pruner-29331360-494rn"] Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.669508 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-jjhb5"] Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.668880 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.668978 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.670395 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.669139 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.669243 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.669370 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.670867 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.669485 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.669807 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.671186 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-tlvk7"] Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.669838 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.671587 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-pruner-29331360-494rn" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.669917 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.669929 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.670087 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.670159 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.671305 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.672023 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-jjhb5" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.672460 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.672900 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fg895"] Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.673100 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-tlvk7" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.690848 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.691152 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.691391 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.692069 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.692285 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fsmpw\" (UniqueName: \"kubernetes.io/projected/95dc52d1-8ca7-45ca-83c4-0eedafd40085-kube-api-access-fsmpw\") pod \"controller-manager-879f6c89f-vkf46\" (UID: \"95dc52d1-8ca7-45ca-83c4-0eedafd40085\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vkf46" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.692323 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-ljjfc"] Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.692363 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/95dc52d1-8ca7-45ca-83c4-0eedafd40085-serving-cert\") pod \"controller-manager-879f6c89f-vkf46\" (UID: \"95dc52d1-8ca7-45ca-83c4-0eedafd40085\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vkf46" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.692404 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/95dc52d1-8ca7-45ca-83c4-0eedafd40085-config\") pod \"controller-manager-879f6c89f-vkf46\" (UID: \"95dc52d1-8ca7-45ca-83c4-0eedafd40085\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vkf46" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.692424 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fg895" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.692456 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/8e759ed2-c526-4ab0-827d-8a7dee2998df-machine-approver-tls\") pod \"machine-approver-56656f9798-5wd5r\" (UID: \"8e759ed2-c526-4ab0-827d-8a7dee2998df\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5wd5r" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.692467 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.692500 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjxsg\" (UniqueName: \"kubernetes.io/projected/2950f912-e1e2-4640-b946-fd9ac7f2a8d3-kube-api-access-bjxsg\") pod \"route-controller-manager-6576b87f9c-v4kxw\" (UID: \"2950f912-e1e2-4640-b946-fd9ac7f2a8d3\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v4kxw" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.692534 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2950f912-e1e2-4640-b946-fd9ac7f2a8d3-config\") pod \"route-controller-manager-6576b87f9c-v4kxw\" (UID: \"2950f912-e1e2-4640-b946-fd9ac7f2a8d3\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v4kxw" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.692580 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mzc2w\" (UniqueName: \"kubernetes.io/projected/8e759ed2-c526-4ab0-827d-8a7dee2998df-kube-api-access-mzc2w\") pod \"machine-approver-56656f9798-5wd5r\" (UID: \"8e759ed2-c526-4ab0-827d-8a7dee2998df\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5wd5r" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.692645 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2950f912-e1e2-4640-b946-fd9ac7f2a8d3-client-ca\") pod \"route-controller-manager-6576b87f9c-v4kxw\" (UID: \"2950f912-e1e2-4640-b946-fd9ac7f2a8d3\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v4kxw" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.692701 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/95dc52d1-8ca7-45ca-83c4-0eedafd40085-client-ca\") pod \"controller-manager-879f6c89f-vkf46\" (UID: \"95dc52d1-8ca7-45ca-83c4-0eedafd40085\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vkf46" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.692736 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e759ed2-c526-4ab0-827d-8a7dee2998df-config\") pod \"machine-approver-56656f9798-5wd5r\" (UID: \"8e759ed2-c526-4ab0-827d-8a7dee2998df\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5wd5r" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.692768 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2950f912-e1e2-4640-b946-fd9ac7f2a8d3-serving-cert\") pod \"route-controller-manager-6576b87f9c-v4kxw\" (UID: \"2950f912-e1e2-4640-b946-fd9ac7f2a8d3\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v4kxw" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.692813 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/95dc52d1-8ca7-45ca-83c4-0eedafd40085-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-vkf46\" (UID: \"95dc52d1-8ca7-45ca-83c4-0eedafd40085\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vkf46" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.692891 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8e759ed2-c526-4ab0-827d-8a7dee2998df-auth-proxy-config\") pod \"machine-approver-56656f9798-5wd5r\" (UID: \"8e759ed2-c526-4ab0-827d-8a7dee2998df\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5wd5r" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.693486 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.694414 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.695111 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.696836 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-ck4pd"] Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.704096 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.712726 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-b5mzq"] Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.714309 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-krqks"] Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.716106 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-krqks" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.717033 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.717156 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.717183 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.726091 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.726839 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"pruner-dockercfg-p7bcw" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.726858 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.727058 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.727170 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.727251 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"serviceca" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.727330 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.727466 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.727551 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.727821 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.727885 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.727959 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.728078 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.728191 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.728927 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-tdqfv"] Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.729324 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-tdqfv" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.730541 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.732614 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-pjlzq"] Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.732838 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.733002 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.733264 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.736856 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.737045 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.737298 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.739365 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-vxdxg"] Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.739686 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ltttx"] Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.739924 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-wxsv2"] Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.740289 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-25pkg"] Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.740364 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.740391 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-wxsv2" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.740403 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-vxdxg" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.740818 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ltttx" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.740921 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.741021 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-8d8x5"] Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.741323 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.741343 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.741551 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.741584 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-pjlzq" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.741756 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-25pkg" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.743811 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.744002 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.744030 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.744310 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.744508 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.744713 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.750965 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-tzhwk"] Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.751620 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-tzhwk" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.754532 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-p9kxf"] Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.755370 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-p9kxf" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.768233 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.770846 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-2ghtr"] Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.771435 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.772014 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2ghtr" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.772825 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n69qz"] Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.773328 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n69qz" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.775012 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-lpxcq"] Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.775902 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-lpxcq" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.778033 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-tmj7c"] Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.780423 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-fqmxp"] Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.781781 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-fqmxp" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.783509 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-zkgkc"] Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.786226 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-zkgkc" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.786294 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.787627 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c4cg2"] Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.791224 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cpnmt"] Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.791404 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c4cg2" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.792326 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4kzvf"] Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.792552 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cpnmt" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.793233 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b5glr"] Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.793313 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2950f912-e1e2-4640-b946-fd9ac7f2a8d3-config\") pod \"route-controller-manager-6576b87f9c-v4kxw\" (UID: \"2950f912-e1e2-4640-b946-fd9ac7f2a8d3\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v4kxw" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.793335 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mzc2w\" (UniqueName: \"kubernetes.io/projected/8e759ed2-c526-4ab0-827d-8a7dee2998df-kube-api-access-mzc2w\") pod \"machine-approver-56656f9798-5wd5r\" (UID: \"8e759ed2-c526-4ab0-827d-8a7dee2998df\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5wd5r" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.793361 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-tmj7c\" (UID: \"d7ec4b5e-d6b4-4681-a231-c7811661c6cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-tmj7c" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.793383 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2950f912-e1e2-4640-b946-fd9ac7f2a8d3-client-ca\") pod \"route-controller-manager-6576b87f9c-v4kxw\" (UID: \"2950f912-e1e2-4640-b946-fd9ac7f2a8d3\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v4kxw" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.793401 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bcxxq\" (UniqueName: \"kubernetes.io/projected/86941755-9852-4757-82b6-c4aa2c3ffa43-kube-api-access-bcxxq\") pod \"openshift-apiserver-operator-796bbdcf4f-p9jtz\" (UID: \"86941755-9852-4757-82b6-c4aa2c3ffa43\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-p9jtz" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.793419 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bb075d42-7ef0-4ad4-b817-bdd611f0d7ed-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-rjwhx\" (UID: \"bb075d42-7ef0-4ad4-b817-bdd611f0d7ed\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rjwhx" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.793434 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d74a2ad0-73e6-4f0d-a804-b7ede2cf8835-config\") pod \"console-operator-58897d9998-tlvk7\" (UID: \"d74a2ad0-73e6-4f0d-a804-b7ede2cf8835\") " pod="openshift-console-operator/console-operator-58897d9998-tlvk7" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.793451 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-tmj7c\" (UID: \"d7ec4b5e-d6b4-4681-a231-c7811661c6cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-tmj7c" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.793467 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/1a14c5f6-9dfd-47fd-acf7-65ec6bbd5149-images\") pod \"machine-api-operator-5694c8668f-ljjfc\" (UID: \"1a14c5f6-9dfd-47fd-acf7-65ec6bbd5149\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-ljjfc" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.793482 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/bb075d42-7ef0-4ad4-b817-bdd611f0d7ed-etcd-client\") pod \"apiserver-7bbb656c7d-rjwhx\" (UID: \"bb075d42-7ef0-4ad4-b817-bdd611f0d7ed\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rjwhx" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.793500 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/95dc52d1-8ca7-45ca-83c4-0eedafd40085-client-ca\") pod \"controller-manager-879f6c89f-vkf46\" (UID: \"95dc52d1-8ca7-45ca-83c4-0eedafd40085\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vkf46" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.793515 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-745vl\" (UniqueName: \"kubernetes.io/projected/71b5b4fb-2e90-4b39-b976-6806ff38ed8e-kube-api-access-745vl\") pod \"apiserver-76f77b778f-ck4pd\" (UID: \"71b5b4fb-2e90-4b39-b976-6806ff38ed8e\") " pod="openshift-apiserver/apiserver-76f77b778f-ck4pd" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.793538 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/86941755-9852-4757-82b6-c4aa2c3ffa43-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-p9jtz\" (UID: \"86941755-9852-4757-82b6-c4aa2c3ffa43\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-p9jtz" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.793553 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lsx2j\" (UniqueName: \"kubernetes.io/projected/1a14c5f6-9dfd-47fd-acf7-65ec6bbd5149-kube-api-access-lsx2j\") pod \"machine-api-operator-5694c8668f-ljjfc\" (UID: \"1a14c5f6-9dfd-47fd-acf7-65ec6bbd5149\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-ljjfc" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.793571 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e759ed2-c526-4ab0-827d-8a7dee2998df-config\") pod \"machine-approver-56656f9798-5wd5r\" (UID: \"8e759ed2-c526-4ab0-827d-8a7dee2998df\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5wd5r" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.793607 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2950f912-e1e2-4640-b946-fd9ac7f2a8d3-serving-cert\") pod \"route-controller-manager-6576b87f9c-v4kxw\" (UID: \"2950f912-e1e2-4640-b946-fd9ac7f2a8d3\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v4kxw" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.793689 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/71b5b4fb-2e90-4b39-b976-6806ff38ed8e-config\") pod \"apiserver-76f77b778f-ck4pd\" (UID: \"71b5b4fb-2e90-4b39-b976-6806ff38ed8e\") " pod="openshift-apiserver/apiserver-76f77b778f-ck4pd" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.793718 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mhcpg\" (UniqueName: \"kubernetes.io/projected/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-kube-api-access-mhcpg\") pod \"oauth-openshift-558db77b4-tmj7c\" (UID: \"d7ec4b5e-d6b4-4681-a231-c7811661c6cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-tmj7c" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.793738 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1a14c5f6-9dfd-47fd-acf7-65ec6bbd5149-config\") pod \"machine-api-operator-5694c8668f-ljjfc\" (UID: \"1a14c5f6-9dfd-47fd-acf7-65ec6bbd5149\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-ljjfc" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.793756 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6801f4d5-5cce-4276-833a-20c81735e3be-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-b5mzq\" (UID: \"6801f4d5-5cce-4276-833a-20c81735e3be\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-b5mzq" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.793791 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/95dc52d1-8ca7-45ca-83c4-0eedafd40085-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-vkf46\" (UID: \"95dc52d1-8ca7-45ca-83c4-0eedafd40085\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vkf46" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.793808 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/87f9f2fd-139e-4ac0-a51c-8473ae6f5986-serving-cert\") pod \"openshift-config-operator-7777fb866f-jjhb5\" (UID: \"87f9f2fd-139e-4ac0-a51c-8473ae6f5986\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-jjhb5" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.794166 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e759ed2-c526-4ab0-827d-8a7dee2998df-config\") pod \"machine-approver-56656f9798-5wd5r\" (UID: \"8e759ed2-c526-4ab0-827d-8a7dee2998df\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5wd5r" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.794258 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/71b5b4fb-2e90-4b39-b976-6806ff38ed8e-etcd-serving-ca\") pod \"apiserver-76f77b778f-ck4pd\" (UID: \"71b5b4fb-2e90-4b39-b976-6806ff38ed8e\") " pod="openshift-apiserver/apiserver-76f77b778f-ck4pd" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.794356 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2950f912-e1e2-4640-b946-fd9ac7f2a8d3-client-ca\") pod \"route-controller-manager-6576b87f9c-v4kxw\" (UID: \"2950f912-e1e2-4640-b946-fd9ac7f2a8d3\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v4kxw" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.794447 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/bb075d42-7ef0-4ad4-b817-bdd611f0d7ed-encryption-config\") pod \"apiserver-7bbb656c7d-rjwhx\" (UID: \"bb075d42-7ef0-4ad4-b817-bdd611f0d7ed\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rjwhx" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.794479 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-tmj7c\" (UID: \"d7ec4b5e-d6b4-4681-a231-c7811661c6cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-tmj7c" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.794532 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8e759ed2-c526-4ab0-827d-8a7dee2998df-auth-proxy-config\") pod \"machine-approver-56656f9798-5wd5r\" (UID: \"8e759ed2-c526-4ab0-827d-8a7dee2998df\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5wd5r" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.794560 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7s4bg\" (UniqueName: \"kubernetes.io/projected/fccf1013-b838-4303-8da2-4861e234540b-kube-api-access-7s4bg\") pod \"openshift-controller-manager-operator-756b6f6bc6-krqks\" (UID: \"fccf1013-b838-4303-8da2-4861e234540b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-krqks" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.794589 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/71b5b4fb-2e90-4b39-b976-6806ff38ed8e-image-import-ca\") pod \"apiserver-76f77b778f-ck4pd\" (UID: \"71b5b4fb-2e90-4b39-b976-6806ff38ed8e\") " pod="openshift-apiserver/apiserver-76f77b778f-ck4pd" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.794616 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/71b5b4fb-2e90-4b39-b976-6806ff38ed8e-trusted-ca-bundle\") pod \"apiserver-76f77b778f-ck4pd\" (UID: \"71b5b4fb-2e90-4b39-b976-6806ff38ed8e\") " pod="openshift-apiserver/apiserver-76f77b778f-ck4pd" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.794639 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/bb075d42-7ef0-4ad4-b817-bdd611f0d7ed-audit-policies\") pod \"apiserver-7bbb656c7d-rjwhx\" (UID: \"bb075d42-7ef0-4ad4-b817-bdd611f0d7ed\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rjwhx" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.794664 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-tmj7c\" (UID: \"d7ec4b5e-d6b4-4681-a231-c7811661c6cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-tmj7c" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.794705 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/71b5b4fb-2e90-4b39-b976-6806ff38ed8e-serving-cert\") pod \"apiserver-76f77b778f-ck4pd\" (UID: \"71b5b4fb-2e90-4b39-b976-6806ff38ed8e\") " pod="openshift-apiserver/apiserver-76f77b778f-ck4pd" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.794729 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dw4fn\" (UniqueName: \"kubernetes.io/projected/d74a2ad0-73e6-4f0d-a804-b7ede2cf8835-kube-api-access-dw4fn\") pod \"console-operator-58897d9998-tlvk7\" (UID: \"d74a2ad0-73e6-4f0d-a804-b7ede2cf8835\") " pod="openshift-console-operator/console-operator-58897d9998-tlvk7" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.794750 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-tmj7c\" (UID: \"d7ec4b5e-d6b4-4681-a231-c7811661c6cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-tmj7c" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.794771 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lzls7\" (UniqueName: \"kubernetes.io/projected/87f9f2fd-139e-4ac0-a51c-8473ae6f5986-kube-api-access-lzls7\") pod \"openshift-config-operator-7777fb866f-jjhb5\" (UID: \"87f9f2fd-139e-4ac0-a51c-8473ae6f5986\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-jjhb5" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.794799 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6801f4d5-5cce-4276-833a-20c81735e3be-serving-cert\") pod \"authentication-operator-69f744f599-b5mzq\" (UID: \"6801f4d5-5cce-4276-833a-20c81735e3be\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-b5mzq" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.794822 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fccf1013-b838-4303-8da2-4861e234540b-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-krqks\" (UID: \"fccf1013-b838-4303-8da2-4861e234540b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-krqks" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.794844 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/71b5b4fb-2e90-4b39-b976-6806ff38ed8e-node-pullsecrets\") pod \"apiserver-76f77b778f-ck4pd\" (UID: \"71b5b4fb-2e90-4b39-b976-6806ff38ed8e\") " pod="openshift-apiserver/apiserver-76f77b778f-ck4pd" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.794874 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fsmpw\" (UniqueName: \"kubernetes.io/projected/95dc52d1-8ca7-45ca-83c4-0eedafd40085-kube-api-access-fsmpw\") pod \"controller-manager-879f6c89f-vkf46\" (UID: \"95dc52d1-8ca7-45ca-83c4-0eedafd40085\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vkf46" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.794910 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/1a14c5f6-9dfd-47fd-acf7-65ec6bbd5149-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-ljjfc\" (UID: \"1a14c5f6-9dfd-47fd-acf7-65ec6bbd5149\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-ljjfc" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.794934 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/87f9f2fd-139e-4ac0-a51c-8473ae6f5986-available-featuregates\") pod \"openshift-config-operator-7777fb866f-jjhb5\" (UID: \"87f9f2fd-139e-4ac0-a51c-8473ae6f5986\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-jjhb5" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.794958 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/95dc52d1-8ca7-45ca-83c4-0eedafd40085-serving-cert\") pod \"controller-manager-879f6c89f-vkf46\" (UID: \"95dc52d1-8ca7-45ca-83c4-0eedafd40085\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vkf46" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.794985 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-tmj7c\" (UID: \"d7ec4b5e-d6b4-4681-a231-c7811661c6cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-tmj7c" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.795011 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b8mt2\" (UniqueName: \"kubernetes.io/projected/6801f4d5-5cce-4276-833a-20c81735e3be-kube-api-access-b8mt2\") pod \"authentication-operator-69f744f599-b5mzq\" (UID: \"6801f4d5-5cce-4276-833a-20c81735e3be\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-b5mzq" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.795038 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/95dc52d1-8ca7-45ca-83c4-0eedafd40085-config\") pod \"controller-manager-879f6c89f-vkf46\" (UID: \"95dc52d1-8ca7-45ca-83c4-0eedafd40085\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vkf46" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.795059 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71b5b4fb-2e90-4b39-b976-6806ff38ed8e-audit-dir\") pod \"apiserver-76f77b778f-ck4pd\" (UID: \"71b5b4fb-2e90-4b39-b976-6806ff38ed8e\") " pod="openshift-apiserver/apiserver-76f77b778f-ck4pd" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.795082 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bb075d42-7ef0-4ad4-b817-bdd611f0d7ed-serving-cert\") pod \"apiserver-7bbb656c7d-rjwhx\" (UID: \"bb075d42-7ef0-4ad4-b817-bdd611f0d7ed\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rjwhx" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.795106 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-tmj7c\" (UID: \"d7ec4b5e-d6b4-4681-a231-c7811661c6cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-tmj7c" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.795128 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-tmj7c\" (UID: \"d7ec4b5e-d6b4-4681-a231-c7811661c6cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-tmj7c" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.795165 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/af03bd84-939d-4a5e-9336-7f53108cdc33-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-fg895\" (UID: \"af03bd84-939d-4a5e-9336-7f53108cdc33\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fg895" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.795202 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/bb075d42-7ef0-4ad4-b817-bdd611f0d7ed-audit-dir\") pod \"apiserver-7bbb656c7d-rjwhx\" (UID: \"bb075d42-7ef0-4ad4-b817-bdd611f0d7ed\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rjwhx" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.795226 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6801f4d5-5cce-4276-833a-20c81735e3be-config\") pod \"authentication-operator-69f744f599-b5mzq\" (UID: \"6801f4d5-5cce-4276-833a-20c81735e3be\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-b5mzq" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.795278 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/8e759ed2-c526-4ab0-827d-8a7dee2998df-machine-approver-tls\") pod \"machine-approver-56656f9798-5wd5r\" (UID: \"8e759ed2-c526-4ab0-827d-8a7dee2998df\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5wd5r" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.795304 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/86941755-9852-4757-82b6-c4aa2c3ffa43-config\") pod \"openshift-apiserver-operator-796bbdcf4f-p9jtz\" (UID: \"86941755-9852-4757-82b6-c4aa2c3ffa43\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-p9jtz" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.795328 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-audit-policies\") pod \"oauth-openshift-558db77b4-tmj7c\" (UID: \"d7ec4b5e-d6b4-4681-a231-c7811661c6cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-tmj7c" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.795351 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-audit-dir\") pod \"oauth-openshift-558db77b4-tmj7c\" (UID: \"d7ec4b5e-d6b4-4681-a231-c7811661c6cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-tmj7c" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.795374 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-tmj7c\" (UID: \"d7ec4b5e-d6b4-4681-a231-c7811661c6cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-tmj7c" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.795397 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-tmj7c\" (UID: \"d7ec4b5e-d6b4-4681-a231-c7811661c6cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-tmj7c" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.795442 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t2th5\" (UniqueName: \"kubernetes.io/projected/84cd9a7a-bb01-43de-9055-deccbc8529b2-kube-api-access-t2th5\") pod \"image-pruner-29331360-494rn\" (UID: \"84cd9a7a-bb01-43de-9055-deccbc8529b2\") " pod="openshift-image-registry/image-pruner-29331360-494rn" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.795463 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/71b5b4fb-2e90-4b39-b976-6806ff38ed8e-etcd-client\") pod \"apiserver-76f77b778f-ck4pd\" (UID: \"71b5b4fb-2e90-4b39-b976-6806ff38ed8e\") " pod="openshift-apiserver/apiserver-76f77b778f-ck4pd" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.795478 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d74a2ad0-73e6-4f0d-a804-b7ede2cf8835-trusted-ca\") pod \"console-operator-58897d9998-tlvk7\" (UID: \"d74a2ad0-73e6-4f0d-a804-b7ede2cf8835\") " pod="openshift-console-operator/console-operator-58897d9998-tlvk7" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.795492 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-tmj7c\" (UID: \"d7ec4b5e-d6b4-4681-a231-c7811661c6cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-tmj7c" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.795508 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/71b5b4fb-2e90-4b39-b976-6806ff38ed8e-audit\") pod \"apiserver-76f77b778f-ck4pd\" (UID: \"71b5b4fb-2e90-4b39-b976-6806ff38ed8e\") " pod="openshift-apiserver/apiserver-76f77b778f-ck4pd" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.795522 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/bb075d42-7ef0-4ad4-b817-bdd611f0d7ed-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-rjwhx\" (UID: \"bb075d42-7ef0-4ad4-b817-bdd611f0d7ed\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rjwhx" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.795537 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6801f4d5-5cce-4276-833a-20c81735e3be-service-ca-bundle\") pod \"authentication-operator-69f744f599-b5mzq\" (UID: \"6801f4d5-5cce-4276-833a-20c81735e3be\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-b5mzq" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.795555 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/84cd9a7a-bb01-43de-9055-deccbc8529b2-serviceca\") pod \"image-pruner-29331360-494rn\" (UID: \"84cd9a7a-bb01-43de-9055-deccbc8529b2\") " pod="openshift-image-registry/image-pruner-29331360-494rn" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.795569 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqw99\" (UniqueName: \"kubernetes.io/projected/fa677745-a987-479c-b60a-2220f1bd58d1-kube-api-access-fqw99\") pod \"downloads-7954f5f757-tdqfv\" (UID: \"fa677745-a987-479c-b60a-2220f1bd58d1\") " pod="openshift-console/downloads-7954f5f757-tdqfv" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.795589 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nrbdr\" (UniqueName: \"kubernetes.io/projected/bb075d42-7ef0-4ad4-b817-bdd611f0d7ed-kube-api-access-nrbdr\") pod \"apiserver-7bbb656c7d-rjwhx\" (UID: \"bb075d42-7ef0-4ad4-b817-bdd611f0d7ed\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rjwhx" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.795603 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d74a2ad0-73e6-4f0d-a804-b7ede2cf8835-serving-cert\") pod \"console-operator-58897d9998-tlvk7\" (UID: \"d74a2ad0-73e6-4f0d-a804-b7ede2cf8835\") " pod="openshift-console-operator/console-operator-58897d9998-tlvk7" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.795619 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9mgd5\" (UniqueName: \"kubernetes.io/projected/af03bd84-939d-4a5e-9336-7f53108cdc33-kube-api-access-9mgd5\") pod \"cluster-samples-operator-665b6dd947-fg895\" (UID: \"af03bd84-939d-4a5e-9336-7f53108cdc33\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fg895" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.795638 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bjxsg\" (UniqueName: \"kubernetes.io/projected/2950f912-e1e2-4640-b946-fd9ac7f2a8d3-kube-api-access-bjxsg\") pod \"route-controller-manager-6576b87f9c-v4kxw\" (UID: \"2950f912-e1e2-4640-b946-fd9ac7f2a8d3\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v4kxw" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.795659 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/71b5b4fb-2e90-4b39-b976-6806ff38ed8e-encryption-config\") pod \"apiserver-76f77b778f-ck4pd\" (UID: \"71b5b4fb-2e90-4b39-b976-6806ff38ed8e\") " pod="openshift-apiserver/apiserver-76f77b778f-ck4pd" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.795718 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fccf1013-b838-4303-8da2-4861e234540b-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-krqks\" (UID: \"fccf1013-b838-4303-8da2-4861e234540b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-krqks" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.795793 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8e759ed2-c526-4ab0-827d-8a7dee2998df-auth-proxy-config\") pod \"machine-approver-56656f9798-5wd5r\" (UID: \"8e759ed2-c526-4ab0-827d-8a7dee2998df\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5wd5r" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.795536 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rsxc6"] Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.796494 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-86lvz"] Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.797714 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-npdfq"] Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.798379 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-jjhb5"] Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.798531 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-npdfq" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.799025 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rsxc6" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.799252 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-86lvz" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.799720 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2950f912-e1e2-4640-b946-fd9ac7f2a8d3-serving-cert\") pod \"route-controller-manager-6576b87f9c-v4kxw\" (UID: \"2950f912-e1e2-4640-b946-fd9ac7f2a8d3\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v4kxw" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.795618 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b5glr" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.795663 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4kzvf" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.800136 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-rjwhx"] Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.801083 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-hpx62"] Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.801357 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/95dc52d1-8ca7-45ca-83c4-0eedafd40085-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-vkf46\" (UID: \"95dc52d1-8ca7-45ca-83c4-0eedafd40085\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vkf46" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.801709 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-hpx62" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.803753 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-wp2tt"] Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.804098 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/8e759ed2-c526-4ab0-827d-8a7dee2998df-machine-approver-tls\") pod \"machine-approver-56656f9798-5wd5r\" (UID: \"8e759ed2-c526-4ab0-827d-8a7dee2998df\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5wd5r" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.804328 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.804450 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wp2tt" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.804985 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2950f912-e1e2-4640-b946-fd9ac7f2a8d3-config\") pod \"route-controller-manager-6576b87f9c-v4kxw\" (UID: \"2950f912-e1e2-4640-b946-fd9ac7f2a8d3\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v4kxw" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.805875 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/95dc52d1-8ca7-45ca-83c4-0eedafd40085-serving-cert\") pod \"controller-manager-879f6c89f-vkf46\" (UID: \"95dc52d1-8ca7-45ca-83c4-0eedafd40085\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vkf46" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.806696 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/95dc52d1-8ca7-45ca-83c4-0eedafd40085-client-ca\") pod \"controller-manager-879f6c89f-vkf46\" (UID: \"95dc52d1-8ca7-45ca-83c4-0eedafd40085\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vkf46" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.806786 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331360-t4jlf"] Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.808059 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/95dc52d1-8ca7-45ca-83c4-0eedafd40085-config\") pod \"controller-manager-879f6c89f-vkf46\" (UID: \"95dc52d1-8ca7-45ca-83c4-0eedafd40085\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vkf46" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.808142 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331360-t4jlf" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.808549 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-pruner-29331360-494rn"] Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.810126 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-p9jtz"] Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.811239 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-kqmk8"] Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.812004 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-kqmk8" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.812288 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-ghwrl"] Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.813156 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-ghwrl" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.814459 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-t65kf"] Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.815555 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-t65kf" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.817316 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-tlvk7"] Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.820398 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-vkf46"] Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.820942 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fg895"] Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.822556 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-tdqfv"] Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.823267 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ltttx"] Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.826322 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-pjlzq"] Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.826370 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-25pkg"] Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.826382 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-8d8x5"] Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.831034 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.831336 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-vxdxg"] Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.837495 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-2ghtr"] Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.841905 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.842150 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c4cg2"] Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.846497 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-zkgkc"] Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.849234 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-krqks"] Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.851493 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cpnmt"] Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.854177 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-wxsv2"] Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.855345 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-npdfq"] Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.857066 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-fqmxp"] Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.859988 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rsxc6"] Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.861844 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.865485 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n69qz"] Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.870772 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-tzhwk"] Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.872057 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-v4kxw"] Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.873253 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-lpxcq"] Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.874478 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4kzvf"] Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.875825 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-ghwrl"] Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.876956 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-fjz5z"] Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.877851 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-fjz5z" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.878156 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-h786z"] Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.878655 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-h786z" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.879375 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-86lvz"] Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.880488 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-kqmk8"] Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.881904 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331360-t4jlf"] Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.881977 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.883557 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-wp2tt"] Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.884866 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-t65kf"] Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.886143 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b5glr"] Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.887402 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-fjz5z"] Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.888540 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-hpx62"] Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.896330 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/bb075d42-7ef0-4ad4-b817-bdd611f0d7ed-etcd-client\") pod \"apiserver-7bbb656c7d-rjwhx\" (UID: \"bb075d42-7ef0-4ad4-b817-bdd611f0d7ed\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rjwhx" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.896366 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0d5e578e-60b8-400c-9545-110600602c3c-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-tzhwk\" (UID: \"0d5e578e-60b8-400c-9545-110600602c3c\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-tzhwk" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.896394 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-745vl\" (UniqueName: \"kubernetes.io/projected/71b5b4fb-2e90-4b39-b976-6806ff38ed8e-kube-api-access-745vl\") pod \"apiserver-76f77b778f-ck4pd\" (UID: \"71b5b4fb-2e90-4b39-b976-6806ff38ed8e\") " pod="openshift-apiserver/apiserver-76f77b778f-ck4pd" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.896413 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/86941755-9852-4757-82b6-c4aa2c3ffa43-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-p9jtz\" (UID: \"86941755-9852-4757-82b6-c4aa2c3ffa43\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-p9jtz" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.896486 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lsx2j\" (UniqueName: \"kubernetes.io/projected/1a14c5f6-9dfd-47fd-acf7-65ec6bbd5149-kube-api-access-lsx2j\") pod \"machine-api-operator-5694c8668f-ljjfc\" (UID: \"1a14c5f6-9dfd-47fd-acf7-65ec6bbd5149\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-ljjfc" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.896532 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/71b5b4fb-2e90-4b39-b976-6806ff38ed8e-config\") pod \"apiserver-76f77b778f-ck4pd\" (UID: \"71b5b4fb-2e90-4b39-b976-6806ff38ed8e\") " pod="openshift-apiserver/apiserver-76f77b778f-ck4pd" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.896598 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mhcpg\" (UniqueName: \"kubernetes.io/projected/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-kube-api-access-mhcpg\") pod \"oauth-openshift-558db77b4-tmj7c\" (UID: \"d7ec4b5e-d6b4-4681-a231-c7811661c6cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-tmj7c" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.896648 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1a14c5f6-9dfd-47fd-acf7-65ec6bbd5149-config\") pod \"machine-api-operator-5694c8668f-ljjfc\" (UID: \"1a14c5f6-9dfd-47fd-acf7-65ec6bbd5149\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-ljjfc" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.896713 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6801f4d5-5cce-4276-833a-20c81735e3be-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-b5mzq\" (UID: \"6801f4d5-5cce-4276-833a-20c81735e3be\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-b5mzq" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.896795 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/87f9f2fd-139e-4ac0-a51c-8473ae6f5986-serving-cert\") pod \"openshift-config-operator-7777fb866f-jjhb5\" (UID: \"87f9f2fd-139e-4ac0-a51c-8473ae6f5986\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-jjhb5" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.896874 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/71b5b4fb-2e90-4b39-b976-6806ff38ed8e-etcd-serving-ca\") pod \"apiserver-76f77b778f-ck4pd\" (UID: \"71b5b4fb-2e90-4b39-b976-6806ff38ed8e\") " pod="openshift-apiserver/apiserver-76f77b778f-ck4pd" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.896902 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/bb075d42-7ef0-4ad4-b817-bdd611f0d7ed-encryption-config\") pod \"apiserver-7bbb656c7d-rjwhx\" (UID: \"bb075d42-7ef0-4ad4-b817-bdd611f0d7ed\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rjwhx" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.896958 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-tmj7c\" (UID: \"d7ec4b5e-d6b4-4681-a231-c7811661c6cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-tmj7c" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.896994 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7s4bg\" (UniqueName: \"kubernetes.io/projected/fccf1013-b838-4303-8da2-4861e234540b-kube-api-access-7s4bg\") pod \"openshift-controller-manager-operator-756b6f6bc6-krqks\" (UID: \"fccf1013-b838-4303-8da2-4861e234540b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-krqks" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.897057 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xtqdk\" (UniqueName: \"kubernetes.io/projected/6a6c3149-2d76-48de-963a-c5ef798a569b-kube-api-access-xtqdk\") pod \"packageserver-d55dfcdfc-rsxc6\" (UID: \"6a6c3149-2d76-48de-963a-c5ef798a569b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rsxc6" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.897111 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jsrp4\" (UniqueName: \"kubernetes.io/projected/3164dad9-48a2-463c-a2c7-c668eec1ac77-kube-api-access-jsrp4\") pod \"dns-operator-744455d44c-wxsv2\" (UID: \"3164dad9-48a2-463c-a2c7-c668eec1ac77\") " pod="openshift-dns-operator/dns-operator-744455d44c-wxsv2" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.897144 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/71b5b4fb-2e90-4b39-b976-6806ff38ed8e-image-import-ca\") pod \"apiserver-76f77b778f-ck4pd\" (UID: \"71b5b4fb-2e90-4b39-b976-6806ff38ed8e\") " pod="openshift-apiserver/apiserver-76f77b778f-ck4pd" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.897197 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/0aec761a-1efe-4817-9bca-cf5961e28ac5-srv-cert\") pod \"catalog-operator-68c6474976-cpnmt\" (UID: \"0aec761a-1efe-4817-9bca-cf5961e28ac5\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cpnmt" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.897231 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/71b5b4fb-2e90-4b39-b976-6806ff38ed8e-trusted-ca-bundle\") pod \"apiserver-76f77b778f-ck4pd\" (UID: \"71b5b4fb-2e90-4b39-b976-6806ff38ed8e\") " pod="openshift-apiserver/apiserver-76f77b778f-ck4pd" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.897290 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/bb075d42-7ef0-4ad4-b817-bdd611f0d7ed-audit-policies\") pod \"apiserver-7bbb656c7d-rjwhx\" (UID: \"bb075d42-7ef0-4ad4-b817-bdd611f0d7ed\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rjwhx" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.897353 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-tmj7c\" (UID: \"d7ec4b5e-d6b4-4681-a231-c7811661c6cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-tmj7c" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.897393 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6a6c3149-2d76-48de-963a-c5ef798a569b-webhook-cert\") pod \"packageserver-d55dfcdfc-rsxc6\" (UID: \"6a6c3149-2d76-48de-963a-c5ef798a569b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rsxc6" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.897454 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/71b5b4fb-2e90-4b39-b976-6806ff38ed8e-serving-cert\") pod \"apiserver-76f77b778f-ck4pd\" (UID: \"71b5b4fb-2e90-4b39-b976-6806ff38ed8e\") " pod="openshift-apiserver/apiserver-76f77b778f-ck4pd" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.897523 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dw4fn\" (UniqueName: \"kubernetes.io/projected/d74a2ad0-73e6-4f0d-a804-b7ede2cf8835-kube-api-access-dw4fn\") pod \"console-operator-58897d9998-tlvk7\" (UID: \"d74a2ad0-73e6-4f0d-a804-b7ede2cf8835\") " pod="openshift-console-operator/console-operator-58897d9998-tlvk7" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.897561 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-tmj7c\" (UID: \"d7ec4b5e-d6b4-4681-a231-c7811661c6cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-tmj7c" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.897621 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lzls7\" (UniqueName: \"kubernetes.io/projected/87f9f2fd-139e-4ac0-a51c-8473ae6f5986-kube-api-access-lzls7\") pod \"openshift-config-operator-7777fb866f-jjhb5\" (UID: \"87f9f2fd-139e-4ac0-a51c-8473ae6f5986\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-jjhb5" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.897717 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f3ea161f-2b22-466c-b49b-9f60e9e3437a-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-ltttx\" (UID: \"f3ea161f-2b22-466c-b49b-9f60e9e3437a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ltttx" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.897763 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fccf1013-b838-4303-8da2-4861e234540b-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-krqks\" (UID: \"fccf1013-b838-4303-8da2-4861e234540b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-krqks" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.897826 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/71b5b4fb-2e90-4b39-b976-6806ff38ed8e-node-pullsecrets\") pod \"apiserver-76f77b778f-ck4pd\" (UID: \"71b5b4fb-2e90-4b39-b976-6806ff38ed8e\") " pod="openshift-apiserver/apiserver-76f77b778f-ck4pd" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.897856 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6801f4d5-5cce-4276-833a-20c81735e3be-serving-cert\") pod \"authentication-operator-69f744f599-b5mzq\" (UID: \"6801f4d5-5cce-4276-833a-20c81735e3be\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-b5mzq" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.897929 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/1a14c5f6-9dfd-47fd-acf7-65ec6bbd5149-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-ljjfc\" (UID: \"1a14c5f6-9dfd-47fd-acf7-65ec6bbd5149\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-ljjfc" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.897958 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/71b5b4fb-2e90-4b39-b976-6806ff38ed8e-config\") pod \"apiserver-76f77b778f-ck4pd\" (UID: \"71b5b4fb-2e90-4b39-b976-6806ff38ed8e\") " pod="openshift-apiserver/apiserver-76f77b778f-ck4pd" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.897965 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/87f9f2fd-139e-4ac0-a51c-8473ae6f5986-available-featuregates\") pod \"openshift-config-operator-7777fb866f-jjhb5\" (UID: \"87f9f2fd-139e-4ac0-a51c-8473ae6f5986\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-jjhb5" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.898010 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/0aec761a-1efe-4817-9bca-cf5961e28ac5-profile-collector-cert\") pod \"catalog-operator-68c6474976-cpnmt\" (UID: \"0aec761a-1efe-4817-9bca-cf5961e28ac5\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cpnmt" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.898071 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-tmj7c\" (UID: \"d7ec4b5e-d6b4-4681-a231-c7811661c6cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-tmj7c" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.898104 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3ea161f-2b22-466c-b49b-9f60e9e3437a-config\") pod \"kube-apiserver-operator-766d6c64bb-ltttx\" (UID: \"f3ea161f-2b22-466c-b49b-9f60e9e3437a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ltttx" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.898139 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b8mt2\" (UniqueName: \"kubernetes.io/projected/6801f4d5-5cce-4276-833a-20c81735e3be-kube-api-access-b8mt2\") pod \"authentication-operator-69f744f599-b5mzq\" (UID: \"6801f4d5-5cce-4276-833a-20c81735e3be\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-b5mzq" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.898161 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1a14c5f6-9dfd-47fd-acf7-65ec6bbd5149-config\") pod \"machine-api-operator-5694c8668f-ljjfc\" (UID: \"1a14c5f6-9dfd-47fd-acf7-65ec6bbd5149\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-ljjfc" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.898177 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71b5b4fb-2e90-4b39-b976-6806ff38ed8e-audit-dir\") pod \"apiserver-76f77b778f-ck4pd\" (UID: \"71b5b4fb-2e90-4b39-b976-6806ff38ed8e\") " pod="openshift-apiserver/apiserver-76f77b778f-ck4pd" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.898221 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71b5b4fb-2e90-4b39-b976-6806ff38ed8e-audit-dir\") pod \"apiserver-76f77b778f-ck4pd\" (UID: \"71b5b4fb-2e90-4b39-b976-6806ff38ed8e\") " pod="openshift-apiserver/apiserver-76f77b778f-ck4pd" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.898239 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bb075d42-7ef0-4ad4-b817-bdd611f0d7ed-serving-cert\") pod \"apiserver-7bbb656c7d-rjwhx\" (UID: \"bb075d42-7ef0-4ad4-b817-bdd611f0d7ed\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rjwhx" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.898518 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/71b5b4fb-2e90-4b39-b976-6806ff38ed8e-etcd-serving-ca\") pod \"apiserver-76f77b778f-ck4pd\" (UID: \"71b5b4fb-2e90-4b39-b976-6806ff38ed8e\") " pod="openshift-apiserver/apiserver-76f77b778f-ck4pd" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.898983 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/71b5b4fb-2e90-4b39-b976-6806ff38ed8e-node-pullsecrets\") pod \"apiserver-76f77b778f-ck4pd\" (UID: \"71b5b4fb-2e90-4b39-b976-6806ff38ed8e\") " pod="openshift-apiserver/apiserver-76f77b778f-ck4pd" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.899416 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-tmj7c\" (UID: \"d7ec4b5e-d6b4-4681-a231-c7811661c6cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-tmj7c" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.899802 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-tmj7c\" (UID: \"d7ec4b5e-d6b4-4681-a231-c7811661c6cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-tmj7c" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.900244 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/87f9f2fd-139e-4ac0-a51c-8473ae6f5986-available-featuregates\") pod \"openshift-config-operator-7777fb866f-jjhb5\" (UID: \"87f9f2fd-139e-4ac0-a51c-8473ae6f5986\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-jjhb5" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.900465 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6a6c3149-2d76-48de-963a-c5ef798a569b-apiservice-cert\") pod \"packageserver-d55dfcdfc-rsxc6\" (UID: \"6a6c3149-2d76-48de-963a-c5ef798a569b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rsxc6" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.900512 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5f033f1c-0cb3-4dcf-ba2b-79593657b841-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-c4cg2\" (UID: \"5f033f1c-0cb3-4dcf-ba2b-79593657b841\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c4cg2" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.900546 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2mcc\" (UniqueName: \"kubernetes.io/projected/5f033f1c-0cb3-4dcf-ba2b-79593657b841-kube-api-access-b2mcc\") pod \"kube-storage-version-migrator-operator-b67b599dd-c4cg2\" (UID: \"5f033f1c-0cb3-4dcf-ba2b-79593657b841\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c4cg2" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.900600 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/af03bd84-939d-4a5e-9336-7f53108cdc33-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-fg895\" (UID: \"af03bd84-939d-4a5e-9336-7f53108cdc33\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fg895" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.900631 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-tmj7c\" (UID: \"d7ec4b5e-d6b4-4681-a231-c7811661c6cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-tmj7c" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.900713 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-tmj7c\" (UID: \"d7ec4b5e-d6b4-4681-a231-c7811661c6cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-tmj7c" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.900853 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/71b5b4fb-2e90-4b39-b976-6806ff38ed8e-trusted-ca-bundle\") pod \"apiserver-76f77b778f-ck4pd\" (UID: \"71b5b4fb-2e90-4b39-b976-6806ff38ed8e\") " pod="openshift-apiserver/apiserver-76f77b778f-ck4pd" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.900935 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/bb075d42-7ef0-4ad4-b817-bdd611f0d7ed-audit-dir\") pod \"apiserver-7bbb656c7d-rjwhx\" (UID: \"bb075d42-7ef0-4ad4-b817-bdd611f0d7ed\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rjwhx" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.900996 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6801f4d5-5cce-4276-833a-20c81735e3be-config\") pod \"authentication-operator-69f744f599-b5mzq\" (UID: \"6801f4d5-5cce-4276-833a-20c81735e3be\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-b5mzq" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.901036 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/bb075d42-7ef0-4ad4-b817-bdd611f0d7ed-audit-dir\") pod \"apiserver-7bbb656c7d-rjwhx\" (UID: \"bb075d42-7ef0-4ad4-b817-bdd611f0d7ed\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rjwhx" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.901150 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fccf1013-b838-4303-8da2-4861e234540b-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-krqks\" (UID: \"fccf1013-b838-4303-8da2-4861e234540b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-krqks" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.901332 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0d5e578e-60b8-400c-9545-110600602c3c-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-tzhwk\" (UID: \"0d5e578e-60b8-400c-9545-110600602c3c\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-tzhwk" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.901446 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/86941755-9852-4757-82b6-c4aa2c3ffa43-config\") pod \"openshift-apiserver-operator-796bbdcf4f-p9jtz\" (UID: \"86941755-9852-4757-82b6-c4aa2c3ffa43\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-p9jtz" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.901494 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-audit-policies\") pod \"oauth-openshift-558db77b4-tmj7c\" (UID: \"d7ec4b5e-d6b4-4681-a231-c7811661c6cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-tmj7c" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.901523 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-audit-dir\") pod \"oauth-openshift-558db77b4-tmj7c\" (UID: \"d7ec4b5e-d6b4-4681-a231-c7811661c6cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-tmj7c" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.901448 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bb075d42-7ef0-4ad4-b817-bdd611f0d7ed-serving-cert\") pod \"apiserver-7bbb656c7d-rjwhx\" (UID: \"bb075d42-7ef0-4ad4-b817-bdd611f0d7ed\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rjwhx" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.901554 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-tmj7c\" (UID: \"d7ec4b5e-d6b4-4681-a231-c7811661c6cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-tmj7c" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.901584 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-tmj7c\" (UID: \"d7ec4b5e-d6b4-4681-a231-c7811661c6cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-tmj7c" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.901620 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t2th5\" (UniqueName: \"kubernetes.io/projected/84cd9a7a-bb01-43de-9055-deccbc8529b2-kube-api-access-t2th5\") pod \"image-pruner-29331360-494rn\" (UID: \"84cd9a7a-bb01-43de-9055-deccbc8529b2\") " pod="openshift-image-registry/image-pruner-29331360-494rn" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.901636 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6801f4d5-5cce-4276-833a-20c81735e3be-config\") pod \"authentication-operator-69f744f599-b5mzq\" (UID: \"6801f4d5-5cce-4276-833a-20c81735e3be\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-b5mzq" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.901654 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/71b5b4fb-2e90-4b39-b976-6806ff38ed8e-etcd-client\") pod \"apiserver-76f77b778f-ck4pd\" (UID: \"71b5b4fb-2e90-4b39-b976-6806ff38ed8e\") " pod="openshift-apiserver/apiserver-76f77b778f-ck4pd" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.901698 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/87f9f2fd-139e-4ac0-a51c-8473ae6f5986-serving-cert\") pod \"openshift-config-operator-7777fb866f-jjhb5\" (UID: \"87f9f2fd-139e-4ac0-a51c-8473ae6f5986\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-jjhb5" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.901736 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d74a2ad0-73e6-4f0d-a804-b7ede2cf8835-trusted-ca\") pod \"console-operator-58897d9998-tlvk7\" (UID: \"d74a2ad0-73e6-4f0d-a804-b7ede2cf8835\") " pod="openshift-console-operator/console-operator-58897d9998-tlvk7" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.901768 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-tmj7c\" (UID: \"d7ec4b5e-d6b4-4681-a231-c7811661c6cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-tmj7c" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.901800 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/71b5b4fb-2e90-4b39-b976-6806ff38ed8e-audit\") pod \"apiserver-76f77b778f-ck4pd\" (UID: \"71b5b4fb-2e90-4b39-b976-6806ff38ed8e\") " pod="openshift-apiserver/apiserver-76f77b778f-ck4pd" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.901827 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/bb075d42-7ef0-4ad4-b817-bdd611f0d7ed-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-rjwhx\" (UID: \"bb075d42-7ef0-4ad4-b817-bdd611f0d7ed\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rjwhx" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.901850 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6801f4d5-5cce-4276-833a-20c81735e3be-service-ca-bundle\") pod \"authentication-operator-69f744f599-b5mzq\" (UID: \"6801f4d5-5cce-4276-833a-20c81735e3be\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-b5mzq" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.901876 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f3ea161f-2b22-466c-b49b-9f60e9e3437a-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-ltttx\" (UID: \"f3ea161f-2b22-466c-b49b-9f60e9e3437a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ltttx" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.901905 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/84cd9a7a-bb01-43de-9055-deccbc8529b2-serviceca\") pod \"image-pruner-29331360-494rn\" (UID: \"84cd9a7a-bb01-43de-9055-deccbc8529b2\") " pod="openshift-image-registry/image-pruner-29331360-494rn" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.901929 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqw99\" (UniqueName: \"kubernetes.io/projected/fa677745-a987-479c-b60a-2220f1bd58d1-kube-api-access-fqw99\") pod \"downloads-7954f5f757-tdqfv\" (UID: \"fa677745-a987-479c-b60a-2220f1bd58d1\") " pod="openshift-console/downloads-7954f5f757-tdqfv" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.901956 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjcsg\" (UniqueName: \"kubernetes.io/projected/0aec761a-1efe-4817-9bca-cf5961e28ac5-kube-api-access-pjcsg\") pod \"catalog-operator-68c6474976-cpnmt\" (UID: \"0aec761a-1efe-4817-9bca-cf5961e28ac5\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cpnmt" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.901988 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nrbdr\" (UniqueName: \"kubernetes.io/projected/bb075d42-7ef0-4ad4-b817-bdd611f0d7ed-kube-api-access-nrbdr\") pod \"apiserver-7bbb656c7d-rjwhx\" (UID: \"bb075d42-7ef0-4ad4-b817-bdd611f0d7ed\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rjwhx" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.902031 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d74a2ad0-73e6-4f0d-a804-b7ede2cf8835-serving-cert\") pod \"console-operator-58897d9998-tlvk7\" (UID: \"d74a2ad0-73e6-4f0d-a804-b7ede2cf8835\") " pod="openshift-console-operator/console-operator-58897d9998-tlvk7" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.902059 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9mgd5\" (UniqueName: \"kubernetes.io/projected/af03bd84-939d-4a5e-9336-7f53108cdc33-kube-api-access-9mgd5\") pod \"cluster-samples-operator-665b6dd947-fg895\" (UID: \"af03bd84-939d-4a5e-9336-7f53108cdc33\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fg895" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.902089 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5f033f1c-0cb3-4dcf-ba2b-79593657b841-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-c4cg2\" (UID: \"5f033f1c-0cb3-4dcf-ba2b-79593657b841\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c4cg2" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.902111 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/6a6c3149-2d76-48de-963a-c5ef798a569b-tmpfs\") pod \"packageserver-d55dfcdfc-rsxc6\" (UID: \"6a6c3149-2d76-48de-963a-c5ef798a569b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rsxc6" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.902138 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d5e578e-60b8-400c-9545-110600602c3c-config\") pod \"kube-controller-manager-operator-78b949d7b-tzhwk\" (UID: \"0d5e578e-60b8-400c-9545-110600602c3c\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-tzhwk" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.902159 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3164dad9-48a2-463c-a2c7-c668eec1ac77-metrics-tls\") pod \"dns-operator-744455d44c-wxsv2\" (UID: \"3164dad9-48a2-463c-a2c7-c668eec1ac77\") " pod="openshift-dns-operator/dns-operator-744455d44c-wxsv2" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.902208 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/86941755-9852-4757-82b6-c4aa2c3ffa43-config\") pod \"openshift-apiserver-operator-796bbdcf4f-p9jtz\" (UID: \"86941755-9852-4757-82b6-c4aa2c3ffa43\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-p9jtz" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.902309 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/71b5b4fb-2e90-4b39-b976-6806ff38ed8e-encryption-config\") pod \"apiserver-76f77b778f-ck4pd\" (UID: \"71b5b4fb-2e90-4b39-b976-6806ff38ed8e\") " pod="openshift-apiserver/apiserver-76f77b778f-ck4pd" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.902321 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-audit-policies\") pod \"oauth-openshift-558db77b4-tmj7c\" (UID: \"d7ec4b5e-d6b4-4681-a231-c7811661c6cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-tmj7c" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.902341 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fccf1013-b838-4303-8da2-4861e234540b-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-krqks\" (UID: \"fccf1013-b838-4303-8da2-4861e234540b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-krqks" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.902428 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-tmj7c\" (UID: \"d7ec4b5e-d6b4-4681-a231-c7811661c6cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-tmj7c" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.902475 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bcxxq\" (UniqueName: \"kubernetes.io/projected/86941755-9852-4757-82b6-c4aa2c3ffa43-kube-api-access-bcxxq\") pod \"openshift-apiserver-operator-796bbdcf4f-p9jtz\" (UID: \"86941755-9852-4757-82b6-c4aa2c3ffa43\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-p9jtz" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.902497 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bb075d42-7ef0-4ad4-b817-bdd611f0d7ed-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-rjwhx\" (UID: \"bb075d42-7ef0-4ad4-b817-bdd611f0d7ed\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rjwhx" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.902522 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-tmj7c\" (UID: \"d7ec4b5e-d6b4-4681-a231-c7811661c6cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-tmj7c" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.902545 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/1a14c5f6-9dfd-47fd-acf7-65ec6bbd5149-images\") pod \"machine-api-operator-5694c8668f-ljjfc\" (UID: \"1a14c5f6-9dfd-47fd-acf7-65ec6bbd5149\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-ljjfc" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.902566 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d74a2ad0-73e6-4f0d-a804-b7ede2cf8835-config\") pod \"console-operator-58897d9998-tlvk7\" (UID: \"d74a2ad0-73e6-4f0d-a804-b7ede2cf8835\") " pod="openshift-console-operator/console-operator-58897d9998-tlvk7" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.902905 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.903430 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-tmj7c\" (UID: \"d7ec4b5e-d6b4-4681-a231-c7811661c6cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-tmj7c" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.903463 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-tmj7c\" (UID: \"d7ec4b5e-d6b4-4681-a231-c7811661c6cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-tmj7c" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.903517 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bb075d42-7ef0-4ad4-b817-bdd611f0d7ed-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-rjwhx\" (UID: \"bb075d42-7ef0-4ad4-b817-bdd611f0d7ed\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rjwhx" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.903530 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d74a2ad0-73e6-4f0d-a804-b7ede2cf8835-config\") pod \"console-operator-58897d9998-tlvk7\" (UID: \"d74a2ad0-73e6-4f0d-a804-b7ede2cf8835\") " pod="openshift-console-operator/console-operator-58897d9998-tlvk7" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.903610 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/71b5b4fb-2e90-4b39-b976-6806ff38ed8e-audit\") pod \"apiserver-76f77b778f-ck4pd\" (UID: \"71b5b4fb-2e90-4b39-b976-6806ff38ed8e\") " pod="openshift-apiserver/apiserver-76f77b778f-ck4pd" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.903632 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-audit-dir\") pod \"oauth-openshift-558db77b4-tmj7c\" (UID: \"d7ec4b5e-d6b4-4681-a231-c7811661c6cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-tmj7c" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.903727 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d74a2ad0-73e6-4f0d-a804-b7ede2cf8835-trusted-ca\") pod \"console-operator-58897d9998-tlvk7\" (UID: \"d74a2ad0-73e6-4f0d-a804-b7ede2cf8835\") " pod="openshift-console-operator/console-operator-58897d9998-tlvk7" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.904220 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/84cd9a7a-bb01-43de-9055-deccbc8529b2-serviceca\") pod \"image-pruner-29331360-494rn\" (UID: \"84cd9a7a-bb01-43de-9055-deccbc8529b2\") " pod="openshift-image-registry/image-pruner-29331360-494rn" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.904321 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/1a14c5f6-9dfd-47fd-acf7-65ec6bbd5149-images\") pod \"machine-api-operator-5694c8668f-ljjfc\" (UID: \"1a14c5f6-9dfd-47fd-acf7-65ec6bbd5149\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-ljjfc" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.904762 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/bb075d42-7ef0-4ad4-b817-bdd611f0d7ed-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-rjwhx\" (UID: \"bb075d42-7ef0-4ad4-b817-bdd611f0d7ed\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rjwhx" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.904877 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6801f4d5-5cce-4276-833a-20c81735e3be-service-ca-bundle\") pod \"authentication-operator-69f744f599-b5mzq\" (UID: \"6801f4d5-5cce-4276-833a-20c81735e3be\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-b5mzq" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.904882 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/71b5b4fb-2e90-4b39-b976-6806ff38ed8e-image-import-ca\") pod \"apiserver-76f77b778f-ck4pd\" (UID: \"71b5b4fb-2e90-4b39-b976-6806ff38ed8e\") " pod="openshift-apiserver/apiserver-76f77b778f-ck4pd" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.904962 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/bb075d42-7ef0-4ad4-b817-bdd611f0d7ed-audit-policies\") pod \"apiserver-7bbb656c7d-rjwhx\" (UID: \"bb075d42-7ef0-4ad4-b817-bdd611f0d7ed\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rjwhx" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.905229 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6801f4d5-5cce-4276-833a-20c81735e3be-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-b5mzq\" (UID: \"6801f4d5-5cce-4276-833a-20c81735e3be\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-b5mzq" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.905798 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-tmj7c\" (UID: \"d7ec4b5e-d6b4-4681-a231-c7811661c6cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-tmj7c" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.906050 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-tmj7c\" (UID: \"d7ec4b5e-d6b4-4681-a231-c7811661c6cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-tmj7c" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.906304 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/bb075d42-7ef0-4ad4-b817-bdd611f0d7ed-etcd-client\") pod \"apiserver-7bbb656c7d-rjwhx\" (UID: \"bb075d42-7ef0-4ad4-b817-bdd611f0d7ed\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rjwhx" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.906535 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d74a2ad0-73e6-4f0d-a804-b7ede2cf8835-serving-cert\") pod \"console-operator-58897d9998-tlvk7\" (UID: \"d74a2ad0-73e6-4f0d-a804-b7ede2cf8835\") " pod="openshift-console-operator/console-operator-58897d9998-tlvk7" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.906547 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/bb075d42-7ef0-4ad4-b817-bdd611f0d7ed-encryption-config\") pod \"apiserver-7bbb656c7d-rjwhx\" (UID: \"bb075d42-7ef0-4ad4-b817-bdd611f0d7ed\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rjwhx" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.907131 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/71b5b4fb-2e90-4b39-b976-6806ff38ed8e-serving-cert\") pod \"apiserver-76f77b778f-ck4pd\" (UID: \"71b5b4fb-2e90-4b39-b976-6806ff38ed8e\") " pod="openshift-apiserver/apiserver-76f77b778f-ck4pd" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.907464 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fccf1013-b838-4303-8da2-4861e234540b-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-krqks\" (UID: \"fccf1013-b838-4303-8da2-4861e234540b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-krqks" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.907617 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/af03bd84-939d-4a5e-9336-7f53108cdc33-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-fg895\" (UID: \"af03bd84-939d-4a5e-9336-7f53108cdc33\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fg895" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.907732 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/1a14c5f6-9dfd-47fd-acf7-65ec6bbd5149-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-ljjfc\" (UID: \"1a14c5f6-9dfd-47fd-acf7-65ec6bbd5149\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-ljjfc" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.908369 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/86941755-9852-4757-82b6-c4aa2c3ffa43-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-p9jtz\" (UID: \"86941755-9852-4757-82b6-c4aa2c3ffa43\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-p9jtz" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.908515 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/71b5b4fb-2e90-4b39-b976-6806ff38ed8e-etcd-client\") pod \"apiserver-76f77b778f-ck4pd\" (UID: \"71b5b4fb-2e90-4b39-b976-6806ff38ed8e\") " pod="openshift-apiserver/apiserver-76f77b778f-ck4pd" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.908787 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-tmj7c\" (UID: \"d7ec4b5e-d6b4-4681-a231-c7811661c6cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-tmj7c" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.909057 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-tmj7c\" (UID: \"d7ec4b5e-d6b4-4681-a231-c7811661c6cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-tmj7c" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.913120 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-tmj7c\" (UID: \"d7ec4b5e-d6b4-4681-a231-c7811661c6cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-tmj7c" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.914985 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-tmj7c\" (UID: \"d7ec4b5e-d6b4-4681-a231-c7811661c6cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-tmj7c" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.915151 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/71b5b4fb-2e90-4b39-b976-6806ff38ed8e-encryption-config\") pod \"apiserver-76f77b778f-ck4pd\" (UID: \"71b5b4fb-2e90-4b39-b976-6806ff38ed8e\") " pod="openshift-apiserver/apiserver-76f77b778f-ck4pd" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.915630 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-tmj7c\" (UID: \"d7ec4b5e-d6b4-4681-a231-c7811661c6cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-tmj7c" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.915729 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6801f4d5-5cce-4276-833a-20c81735e3be-serving-cert\") pod \"authentication-operator-69f744f599-b5mzq\" (UID: \"6801f4d5-5cce-4276-833a-20c81735e3be\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-b5mzq" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.922144 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.942577 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.962549 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Oct 08 00:08:39 crc kubenswrapper[4967]: I1008 00:08:39.989954 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Oct 08 00:08:40 crc kubenswrapper[4967]: I1008 00:08:40.002018 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Oct 08 00:08:40 crc kubenswrapper[4967]: I1008 00:08:40.003642 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0d5e578e-60b8-400c-9545-110600602c3c-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-tzhwk\" (UID: \"0d5e578e-60b8-400c-9545-110600602c3c\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-tzhwk" Oct 08 00:08:40 crc kubenswrapper[4967]: I1008 00:08:40.003760 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xtqdk\" (UniqueName: \"kubernetes.io/projected/6a6c3149-2d76-48de-963a-c5ef798a569b-kube-api-access-xtqdk\") pod \"packageserver-d55dfcdfc-rsxc6\" (UID: \"6a6c3149-2d76-48de-963a-c5ef798a569b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rsxc6" Oct 08 00:08:40 crc kubenswrapper[4967]: I1008 00:08:40.003795 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jsrp4\" (UniqueName: \"kubernetes.io/projected/3164dad9-48a2-463c-a2c7-c668eec1ac77-kube-api-access-jsrp4\") pod \"dns-operator-744455d44c-wxsv2\" (UID: \"3164dad9-48a2-463c-a2c7-c668eec1ac77\") " pod="openshift-dns-operator/dns-operator-744455d44c-wxsv2" Oct 08 00:08:40 crc kubenswrapper[4967]: I1008 00:08:40.003817 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/0aec761a-1efe-4817-9bca-cf5961e28ac5-srv-cert\") pod \"catalog-operator-68c6474976-cpnmt\" (UID: \"0aec761a-1efe-4817-9bca-cf5961e28ac5\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cpnmt" Oct 08 00:08:40 crc kubenswrapper[4967]: I1008 00:08:40.003839 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6a6c3149-2d76-48de-963a-c5ef798a569b-webhook-cert\") pod \"packageserver-d55dfcdfc-rsxc6\" (UID: \"6a6c3149-2d76-48de-963a-c5ef798a569b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rsxc6" Oct 08 00:08:40 crc kubenswrapper[4967]: I1008 00:08:40.003875 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f3ea161f-2b22-466c-b49b-9f60e9e3437a-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-ltttx\" (UID: \"f3ea161f-2b22-466c-b49b-9f60e9e3437a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ltttx" Oct 08 00:08:40 crc kubenswrapper[4967]: I1008 00:08:40.003915 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/0aec761a-1efe-4817-9bca-cf5961e28ac5-profile-collector-cert\") pod \"catalog-operator-68c6474976-cpnmt\" (UID: \"0aec761a-1efe-4817-9bca-cf5961e28ac5\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cpnmt" Oct 08 00:08:40 crc kubenswrapper[4967]: I1008 00:08:40.003939 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3ea161f-2b22-466c-b49b-9f60e9e3437a-config\") pod \"kube-apiserver-operator-766d6c64bb-ltttx\" (UID: \"f3ea161f-2b22-466c-b49b-9f60e9e3437a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ltttx" Oct 08 00:08:40 crc kubenswrapper[4967]: I1008 00:08:40.003985 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6a6c3149-2d76-48de-963a-c5ef798a569b-apiservice-cert\") pod \"packageserver-d55dfcdfc-rsxc6\" (UID: \"6a6c3149-2d76-48de-963a-c5ef798a569b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rsxc6" Oct 08 00:08:40 crc kubenswrapper[4967]: I1008 00:08:40.004011 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5f033f1c-0cb3-4dcf-ba2b-79593657b841-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-c4cg2\" (UID: \"5f033f1c-0cb3-4dcf-ba2b-79593657b841\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c4cg2" Oct 08 00:08:40 crc kubenswrapper[4967]: I1008 00:08:40.004036 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2mcc\" (UniqueName: \"kubernetes.io/projected/5f033f1c-0cb3-4dcf-ba2b-79593657b841-kube-api-access-b2mcc\") pod \"kube-storage-version-migrator-operator-b67b599dd-c4cg2\" (UID: \"5f033f1c-0cb3-4dcf-ba2b-79593657b841\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c4cg2" Oct 08 00:08:40 crc kubenswrapper[4967]: I1008 00:08:40.004068 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0d5e578e-60b8-400c-9545-110600602c3c-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-tzhwk\" (UID: \"0d5e578e-60b8-400c-9545-110600602c3c\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-tzhwk" Oct 08 00:08:40 crc kubenswrapper[4967]: I1008 00:08:40.004106 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f3ea161f-2b22-466c-b49b-9f60e9e3437a-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-ltttx\" (UID: \"f3ea161f-2b22-466c-b49b-9f60e9e3437a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ltttx" Oct 08 00:08:40 crc kubenswrapper[4967]: I1008 00:08:40.004140 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjcsg\" (UniqueName: \"kubernetes.io/projected/0aec761a-1efe-4817-9bca-cf5961e28ac5-kube-api-access-pjcsg\") pod \"catalog-operator-68c6474976-cpnmt\" (UID: \"0aec761a-1efe-4817-9bca-cf5961e28ac5\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cpnmt" Oct 08 00:08:40 crc kubenswrapper[4967]: I1008 00:08:40.004180 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5f033f1c-0cb3-4dcf-ba2b-79593657b841-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-c4cg2\" (UID: \"5f033f1c-0cb3-4dcf-ba2b-79593657b841\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c4cg2" Oct 08 00:08:40 crc kubenswrapper[4967]: I1008 00:08:40.004205 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/6a6c3149-2d76-48de-963a-c5ef798a569b-tmpfs\") pod \"packageserver-d55dfcdfc-rsxc6\" (UID: \"6a6c3149-2d76-48de-963a-c5ef798a569b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rsxc6" Oct 08 00:08:40 crc kubenswrapper[4967]: I1008 00:08:40.004232 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d5e578e-60b8-400c-9545-110600602c3c-config\") pod \"kube-controller-manager-operator-78b949d7b-tzhwk\" (UID: \"0d5e578e-60b8-400c-9545-110600602c3c\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-tzhwk" Oct 08 00:08:40 crc kubenswrapper[4967]: I1008 00:08:40.004255 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3164dad9-48a2-463c-a2c7-c668eec1ac77-metrics-tls\") pod \"dns-operator-744455d44c-wxsv2\" (UID: \"3164dad9-48a2-463c-a2c7-c668eec1ac77\") " pod="openshift-dns-operator/dns-operator-744455d44c-wxsv2" Oct 08 00:08:40 crc kubenswrapper[4967]: I1008 00:08:40.004685 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3ea161f-2b22-466c-b49b-9f60e9e3437a-config\") pod \"kube-apiserver-operator-766d6c64bb-ltttx\" (UID: \"f3ea161f-2b22-466c-b49b-9f60e9e3437a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ltttx" Oct 08 00:08:40 crc kubenswrapper[4967]: I1008 00:08:40.004728 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/6a6c3149-2d76-48de-963a-c5ef798a569b-tmpfs\") pod \"packageserver-d55dfcdfc-rsxc6\" (UID: \"6a6c3149-2d76-48de-963a-c5ef798a569b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rsxc6" Oct 08 00:08:40 crc kubenswrapper[4967]: I1008 00:08:40.006573 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f3ea161f-2b22-466c-b49b-9f60e9e3437a-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-ltttx\" (UID: \"f3ea161f-2b22-466c-b49b-9f60e9e3437a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ltttx" Oct 08 00:08:40 crc kubenswrapper[4967]: I1008 00:08:40.008001 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3164dad9-48a2-463c-a2c7-c668eec1ac77-metrics-tls\") pod \"dns-operator-744455d44c-wxsv2\" (UID: \"3164dad9-48a2-463c-a2c7-c668eec1ac77\") " pod="openshift-dns-operator/dns-operator-744455d44c-wxsv2" Oct 08 00:08:40 crc kubenswrapper[4967]: I1008 00:08:40.022828 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Oct 08 00:08:40 crc kubenswrapper[4967]: I1008 00:08:40.043095 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Oct 08 00:08:40 crc kubenswrapper[4967]: I1008 00:08:40.062111 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Oct 08 00:08:40 crc kubenswrapper[4967]: I1008 00:08:40.102797 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Oct 08 00:08:40 crc kubenswrapper[4967]: I1008 00:08:40.116948 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 00:08:40 crc kubenswrapper[4967]: I1008 00:08:40.123275 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Oct 08 00:08:40 crc kubenswrapper[4967]: I1008 00:08:40.142405 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Oct 08 00:08:40 crc kubenswrapper[4967]: I1008 00:08:40.147180 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0d5e578e-60b8-400c-9545-110600602c3c-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-tzhwk\" (UID: \"0d5e578e-60b8-400c-9545-110600602c3c\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-tzhwk" Oct 08 00:08:40 crc kubenswrapper[4967]: I1008 00:08:40.162954 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Oct 08 00:08:40 crc kubenswrapper[4967]: I1008 00:08:40.166241 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d5e578e-60b8-400c-9545-110600602c3c-config\") pod \"kube-controller-manager-operator-78b949d7b-tzhwk\" (UID: \"0d5e578e-60b8-400c-9545-110600602c3c\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-tzhwk" Oct 08 00:08:40 crc kubenswrapper[4967]: I1008 00:08:40.182860 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Oct 08 00:08:40 crc kubenswrapper[4967]: I1008 00:08:40.202862 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Oct 08 00:08:40 crc kubenswrapper[4967]: I1008 00:08:40.222458 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Oct 08 00:08:40 crc kubenswrapper[4967]: I1008 00:08:40.242459 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Oct 08 00:08:40 crc kubenswrapper[4967]: I1008 00:08:40.262776 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Oct 08 00:08:40 crc kubenswrapper[4967]: I1008 00:08:40.282742 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Oct 08 00:08:40 crc kubenswrapper[4967]: I1008 00:08:40.303002 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Oct 08 00:08:40 crc kubenswrapper[4967]: I1008 00:08:40.324490 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Oct 08 00:08:40 crc kubenswrapper[4967]: I1008 00:08:40.343912 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Oct 08 00:08:40 crc kubenswrapper[4967]: I1008 00:08:40.362202 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Oct 08 00:08:40 crc kubenswrapper[4967]: I1008 00:08:40.383326 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Oct 08 00:08:40 crc kubenswrapper[4967]: I1008 00:08:40.403183 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Oct 08 00:08:40 crc kubenswrapper[4967]: I1008 00:08:40.423491 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Oct 08 00:08:40 crc kubenswrapper[4967]: I1008 00:08:40.442876 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Oct 08 00:08:40 crc kubenswrapper[4967]: I1008 00:08:40.463163 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Oct 08 00:08:40 crc kubenswrapper[4967]: I1008 00:08:40.483783 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Oct 08 00:08:40 crc kubenswrapper[4967]: I1008 00:08:40.502634 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Oct 08 00:08:40 crc kubenswrapper[4967]: I1008 00:08:40.523131 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Oct 08 00:08:40 crc kubenswrapper[4967]: I1008 00:08:40.543296 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Oct 08 00:08:40 crc kubenswrapper[4967]: I1008 00:08:40.562872 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Oct 08 00:08:40 crc kubenswrapper[4967]: I1008 00:08:40.582927 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Oct 08 00:08:40 crc kubenswrapper[4967]: I1008 00:08:40.603230 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Oct 08 00:08:40 crc kubenswrapper[4967]: I1008 00:08:40.623097 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Oct 08 00:08:40 crc kubenswrapper[4967]: I1008 00:08:40.643445 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Oct 08 00:08:40 crc kubenswrapper[4967]: I1008 00:08:40.663199 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Oct 08 00:08:40 crc kubenswrapper[4967]: I1008 00:08:40.683615 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Oct 08 00:08:40 crc kubenswrapper[4967]: I1008 00:08:40.703173 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Oct 08 00:08:40 crc kubenswrapper[4967]: I1008 00:08:40.723583 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Oct 08 00:08:40 crc kubenswrapper[4967]: I1008 00:08:40.743810 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Oct 08 00:08:40 crc kubenswrapper[4967]: I1008 00:08:40.746315 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5f033f1c-0cb3-4dcf-ba2b-79593657b841-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-c4cg2\" (UID: \"5f033f1c-0cb3-4dcf-ba2b-79593657b841\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c4cg2" Oct 08 00:08:40 crc kubenswrapper[4967]: I1008 00:08:40.763782 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Oct 08 00:08:40 crc kubenswrapper[4967]: I1008 00:08:40.783455 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Oct 08 00:08:40 crc kubenswrapper[4967]: I1008 00:08:40.800497 4967 request.go:700] Waited for 1.008533515s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-storage-version-migrator-operator/secrets?fieldSelector=metadata.name%3Dserving-cert&limit=500&resourceVersion=0 Oct 08 00:08:40 crc kubenswrapper[4967]: I1008 00:08:40.802594 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Oct 08 00:08:40 crc kubenswrapper[4967]: I1008 00:08:40.808635 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5f033f1c-0cb3-4dcf-ba2b-79593657b841-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-c4cg2\" (UID: \"5f033f1c-0cb3-4dcf-ba2b-79593657b841\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c4cg2" Oct 08 00:08:40 crc kubenswrapper[4967]: I1008 00:08:40.822969 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Oct 08 00:08:40 crc kubenswrapper[4967]: I1008 00:08:40.843146 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Oct 08 00:08:40 crc kubenswrapper[4967]: I1008 00:08:40.847958 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/0aec761a-1efe-4817-9bca-cf5961e28ac5-srv-cert\") pod \"catalog-operator-68c6474976-cpnmt\" (UID: \"0aec761a-1efe-4817-9bca-cf5961e28ac5\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cpnmt" Oct 08 00:08:40 crc kubenswrapper[4967]: I1008 00:08:40.864625 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Oct 08 00:08:40 crc kubenswrapper[4967]: I1008 00:08:40.883720 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Oct 08 00:08:40 crc kubenswrapper[4967]: I1008 00:08:40.888662 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/0aec761a-1efe-4817-9bca-cf5961e28ac5-profile-collector-cert\") pod \"catalog-operator-68c6474976-cpnmt\" (UID: \"0aec761a-1efe-4817-9bca-cf5961e28ac5\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cpnmt" Oct 08 00:08:40 crc kubenswrapper[4967]: I1008 00:08:40.903782 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Oct 08 00:08:40 crc kubenswrapper[4967]: I1008 00:08:40.943046 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mzc2w\" (UniqueName: \"kubernetes.io/projected/8e759ed2-c526-4ab0-827d-8a7dee2998df-kube-api-access-mzc2w\") pod \"machine-approver-56656f9798-5wd5r\" (UID: \"8e759ed2-c526-4ab0-827d-8a7dee2998df\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5wd5r" Oct 08 00:08:40 crc kubenswrapper[4967]: I1008 00:08:40.953757 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Oct 08 00:08:40 crc kubenswrapper[4967]: I1008 00:08:40.963092 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Oct 08 00:08:40 crc kubenswrapper[4967]: I1008 00:08:40.968444 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6a6c3149-2d76-48de-963a-c5ef798a569b-webhook-cert\") pod \"packageserver-d55dfcdfc-rsxc6\" (UID: \"6a6c3149-2d76-48de-963a-c5ef798a569b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rsxc6" Oct 08 00:08:40 crc kubenswrapper[4967]: I1008 00:08:40.969350 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6a6c3149-2d76-48de-963a-c5ef798a569b-apiservice-cert\") pod \"packageserver-d55dfcdfc-rsxc6\" (UID: \"6a6c3149-2d76-48de-963a-c5ef798a569b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rsxc6" Oct 08 00:08:40 crc kubenswrapper[4967]: I1008 00:08:40.983527 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Oct 08 00:08:41 crc kubenswrapper[4967]: I1008 00:08:41.004130 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Oct 08 00:08:41 crc kubenswrapper[4967]: I1008 00:08:41.022570 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Oct 08 00:08:41 crc kubenswrapper[4967]: I1008 00:08:41.042720 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Oct 08 00:08:41 crc kubenswrapper[4967]: I1008 00:08:41.064296 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Oct 08 00:08:41 crc kubenswrapper[4967]: I1008 00:08:41.083384 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Oct 08 00:08:41 crc kubenswrapper[4967]: I1008 00:08:41.102748 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Oct 08 00:08:41 crc kubenswrapper[4967]: I1008 00:08:41.117662 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 00:08:41 crc kubenswrapper[4967]: I1008 00:08:41.117923 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2h7z8" Oct 08 00:08:41 crc kubenswrapper[4967]: I1008 00:08:41.117996 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 00:08:41 crc kubenswrapper[4967]: I1008 00:08:41.151866 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bjxsg\" (UniqueName: \"kubernetes.io/projected/2950f912-e1e2-4640-b946-fd9ac7f2a8d3-kube-api-access-bjxsg\") pod \"route-controller-manager-6576b87f9c-v4kxw\" (UID: \"2950f912-e1e2-4640-b946-fd9ac7f2a8d3\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v4kxw" Oct 08 00:08:41 crc kubenswrapper[4967]: I1008 00:08:41.163523 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v4kxw" Oct 08 00:08:41 crc kubenswrapper[4967]: I1008 00:08:41.163526 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Oct 08 00:08:41 crc kubenswrapper[4967]: I1008 00:08:41.171042 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fsmpw\" (UniqueName: \"kubernetes.io/projected/95dc52d1-8ca7-45ca-83c4-0eedafd40085-kube-api-access-fsmpw\") pod \"controller-manager-879f6c89f-vkf46\" (UID: \"95dc52d1-8ca7-45ca-83c4-0eedafd40085\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vkf46" Oct 08 00:08:41 crc kubenswrapper[4967]: I1008 00:08:41.184498 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Oct 08 00:08:41 crc kubenswrapper[4967]: I1008 00:08:41.191741 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5wd5r" Oct 08 00:08:41 crc kubenswrapper[4967]: I1008 00:08:41.205003 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Oct 08 00:08:41 crc kubenswrapper[4967]: I1008 00:08:41.224047 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Oct 08 00:08:41 crc kubenswrapper[4967]: I1008 00:08:41.242728 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Oct 08 00:08:41 crc kubenswrapper[4967]: I1008 00:08:41.263282 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Oct 08 00:08:41 crc kubenswrapper[4967]: I1008 00:08:41.283763 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Oct 08 00:08:41 crc kubenswrapper[4967]: I1008 00:08:41.303728 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Oct 08 00:08:41 crc kubenswrapper[4967]: I1008 00:08:41.336317 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Oct 08 00:08:41 crc kubenswrapper[4967]: I1008 00:08:41.342457 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Oct 08 00:08:41 crc kubenswrapper[4967]: I1008 00:08:41.363942 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 08 00:08:41 crc kubenswrapper[4967]: I1008 00:08:41.384649 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 08 00:08:41 crc kubenswrapper[4967]: I1008 00:08:41.402874 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Oct 08 00:08:41 crc kubenswrapper[4967]: I1008 00:08:41.407364 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-v4kxw"] Oct 08 00:08:41 crc kubenswrapper[4967]: I1008 00:08:41.422808 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Oct 08 00:08:41 crc kubenswrapper[4967]: I1008 00:08:41.443084 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Oct 08 00:08:41 crc kubenswrapper[4967]: I1008 00:08:41.446219 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-vkf46" Oct 08 00:08:41 crc kubenswrapper[4967]: I1008 00:08:41.462340 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Oct 08 00:08:41 crc kubenswrapper[4967]: I1008 00:08:41.484120 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Oct 08 00:08:41 crc kubenswrapper[4967]: I1008 00:08:41.510369 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Oct 08 00:08:41 crc kubenswrapper[4967]: I1008 00:08:41.523375 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Oct 08 00:08:41 crc kubenswrapper[4967]: I1008 00:08:41.543013 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Oct 08 00:08:41 crc kubenswrapper[4967]: I1008 00:08:41.567057 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Oct 08 00:08:41 crc kubenswrapper[4967]: I1008 00:08:41.582103 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Oct 08 00:08:41 crc kubenswrapper[4967]: I1008 00:08:41.603285 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Oct 08 00:08:41 crc kubenswrapper[4967]: I1008 00:08:41.623015 4967 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Oct 08 00:08:41 crc kubenswrapper[4967]: I1008 00:08:41.631218 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-vkf46"] Oct 08 00:08:41 crc kubenswrapper[4967]: W1008 00:08:41.638916 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod95dc52d1_8ca7_45ca_83c4_0eedafd40085.slice/crio-17c7e7c8e033ebbfecd21ef1a5396727c0b3abcdb20bfefac115ede13cbfc25f WatchSource:0}: Error finding container 17c7e7c8e033ebbfecd21ef1a5396727c0b3abcdb20bfefac115ede13cbfc25f: Status 404 returned error can't find the container with id 17c7e7c8e033ebbfecd21ef1a5396727c0b3abcdb20bfefac115ede13cbfc25f Oct 08 00:08:41 crc kubenswrapper[4967]: I1008 00:08:41.643000 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Oct 08 00:08:41 crc kubenswrapper[4967]: I1008 00:08:41.682715 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Oct 08 00:08:41 crc kubenswrapper[4967]: I1008 00:08:41.703305 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Oct 08 00:08:41 crc kubenswrapper[4967]: I1008 00:08:41.722161 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Oct 08 00:08:41 crc kubenswrapper[4967]: I1008 00:08:41.742871 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Oct 08 00:08:41 crc kubenswrapper[4967]: I1008 00:08:41.763272 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Oct 08 00:08:41 crc kubenswrapper[4967]: I1008 00:08:41.783208 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Oct 08 00:08:41 crc kubenswrapper[4967]: I1008 00:08:41.800729 4967 request.go:700] Waited for 1.904127338s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-apiserver/serviceaccounts/openshift-apiserver-sa/token Oct 08 00:08:41 crc kubenswrapper[4967]: I1008 00:08:41.818475 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5wd5r" event={"ID":"8e759ed2-c526-4ab0-827d-8a7dee2998df","Type":"ContainerStarted","Data":"6209d492f8e5e9ca1c5fcc3d505e18d5f67917af20f0a825af75be99e4916c44"} Oct 08 00:08:41 crc kubenswrapper[4967]: I1008 00:08:41.818780 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5wd5r" event={"ID":"8e759ed2-c526-4ab0-827d-8a7dee2998df","Type":"ContainerStarted","Data":"c1d3279b983d0686d4c4cabe2b2e4defd686aefab6098e20115c6da5d08c130b"} Oct 08 00:08:41 crc kubenswrapper[4967]: I1008 00:08:41.818811 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5wd5r" event={"ID":"8e759ed2-c526-4ab0-827d-8a7dee2998df","Type":"ContainerStarted","Data":"8912cdc0202696764a34b2c9cbcf7de5d7b0ebbae433fe5c7ac4e69d2515c440"} Oct 08 00:08:41 crc kubenswrapper[4967]: I1008 00:08:41.820621 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v4kxw" event={"ID":"2950f912-e1e2-4640-b946-fd9ac7f2a8d3","Type":"ContainerStarted","Data":"4388470ff7f5a8b15c206786fa0b888701511b6432b249e45dcbe25207ebf21c"} Oct 08 00:08:41 crc kubenswrapper[4967]: I1008 00:08:41.820705 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v4kxw" event={"ID":"2950f912-e1e2-4640-b946-fd9ac7f2a8d3","Type":"ContainerStarted","Data":"5b0455f52fa7244be50a4ba95d0f95417f00483fea24320be312e4c8d08c2a77"} Oct 08 00:08:41 crc kubenswrapper[4967]: I1008 00:08:41.820982 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v4kxw" Oct 08 00:08:41 crc kubenswrapper[4967]: I1008 00:08:41.824838 4967 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-v4kxw container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.33:8443/healthz\": dial tcp 10.217.0.33:8443: connect: connection refused" start-of-body= Oct 08 00:08:41 crc kubenswrapper[4967]: I1008 00:08:41.824906 4967 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v4kxw" podUID="2950f912-e1e2-4640-b946-fd9ac7f2a8d3" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.33:8443/healthz\": dial tcp 10.217.0.33:8443: connect: connection refused" Oct 08 00:08:41 crc kubenswrapper[4967]: I1008 00:08:41.826984 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-745vl\" (UniqueName: \"kubernetes.io/projected/71b5b4fb-2e90-4b39-b976-6806ff38ed8e-kube-api-access-745vl\") pod \"apiserver-76f77b778f-ck4pd\" (UID: \"71b5b4fb-2e90-4b39-b976-6806ff38ed8e\") " pod="openshift-apiserver/apiserver-76f77b778f-ck4pd" Oct 08 00:08:41 crc kubenswrapper[4967]: I1008 00:08:41.828021 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-vkf46" event={"ID":"95dc52d1-8ca7-45ca-83c4-0eedafd40085","Type":"ContainerStarted","Data":"4106e7ce0e367c0ecf5bf205d01584d073e67ff15fcc1437538cbd2d462cc931"} Oct 08 00:08:41 crc kubenswrapper[4967]: I1008 00:08:41.828056 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-vkf46" event={"ID":"95dc52d1-8ca7-45ca-83c4-0eedafd40085","Type":"ContainerStarted","Data":"17c7e7c8e033ebbfecd21ef1a5396727c0b3abcdb20bfefac115ede13cbfc25f"} Oct 08 00:08:41 crc kubenswrapper[4967]: I1008 00:08:41.828291 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-vkf46" Oct 08 00:08:41 crc kubenswrapper[4967]: I1008 00:08:41.835046 4967 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-vkf46 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Oct 08 00:08:41 crc kubenswrapper[4967]: I1008 00:08:41.835097 4967 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-vkf46" podUID="95dc52d1-8ca7-45ca-83c4-0eedafd40085" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" Oct 08 00:08:41 crc kubenswrapper[4967]: I1008 00:08:41.843120 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lsx2j\" (UniqueName: \"kubernetes.io/projected/1a14c5f6-9dfd-47fd-acf7-65ec6bbd5149-kube-api-access-lsx2j\") pod \"machine-api-operator-5694c8668f-ljjfc\" (UID: \"1a14c5f6-9dfd-47fd-acf7-65ec6bbd5149\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-ljjfc" Oct 08 00:08:41 crc kubenswrapper[4967]: I1008 00:08:41.858721 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mhcpg\" (UniqueName: \"kubernetes.io/projected/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-kube-api-access-mhcpg\") pod \"oauth-openshift-558db77b4-tmj7c\" (UID: \"d7ec4b5e-d6b4-4681-a231-c7811661c6cb\") " pod="openshift-authentication/oauth-openshift-558db77b4-tmj7c" Oct 08 00:08:41 crc kubenswrapper[4967]: I1008 00:08:41.887820 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dw4fn\" (UniqueName: \"kubernetes.io/projected/d74a2ad0-73e6-4f0d-a804-b7ede2cf8835-kube-api-access-dw4fn\") pod \"console-operator-58897d9998-tlvk7\" (UID: \"d74a2ad0-73e6-4f0d-a804-b7ede2cf8835\") " pod="openshift-console-operator/console-operator-58897d9998-tlvk7" Oct 08 00:08:41 crc kubenswrapper[4967]: I1008 00:08:41.890399 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-tmj7c" Oct 08 00:08:41 crc kubenswrapper[4967]: I1008 00:08:41.908540 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7s4bg\" (UniqueName: \"kubernetes.io/projected/fccf1013-b838-4303-8da2-4861e234540b-kube-api-access-7s4bg\") pod \"openshift-controller-manager-operator-756b6f6bc6-krqks\" (UID: \"fccf1013-b838-4303-8da2-4861e234540b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-krqks" Oct 08 00:08:41 crc kubenswrapper[4967]: I1008 00:08:41.923532 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b8mt2\" (UniqueName: \"kubernetes.io/projected/6801f4d5-5cce-4276-833a-20c81735e3be-kube-api-access-b8mt2\") pod \"authentication-operator-69f744f599-b5mzq\" (UID: \"6801f4d5-5cce-4276-833a-20c81735e3be\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-b5mzq" Oct 08 00:08:41 crc kubenswrapper[4967]: I1008 00:08:41.936909 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lzls7\" (UniqueName: \"kubernetes.io/projected/87f9f2fd-139e-4ac0-a51c-8473ae6f5986-kube-api-access-lzls7\") pod \"openshift-config-operator-7777fb866f-jjhb5\" (UID: \"87f9f2fd-139e-4ac0-a51c-8473ae6f5986\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-jjhb5" Oct 08 00:08:41 crc kubenswrapper[4967]: I1008 00:08:41.940691 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-jjhb5" Oct 08 00:08:41 crc kubenswrapper[4967]: I1008 00:08:41.949012 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-tlvk7" Oct 08 00:08:41 crc kubenswrapper[4967]: I1008 00:08:41.956149 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqw99\" (UniqueName: \"kubernetes.io/projected/fa677745-a987-479c-b60a-2220f1bd58d1-kube-api-access-fqw99\") pod \"downloads-7954f5f757-tdqfv\" (UID: \"fa677745-a987-479c-b60a-2220f1bd58d1\") " pod="openshift-console/downloads-7954f5f757-tdqfv" Oct 08 00:08:41 crc kubenswrapper[4967]: I1008 00:08:41.961932 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-krqks" Oct 08 00:08:41 crc kubenswrapper[4967]: I1008 00:08:41.971193 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-tdqfv" Oct 08 00:08:41 crc kubenswrapper[4967]: I1008 00:08:41.978456 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t2th5\" (UniqueName: \"kubernetes.io/projected/84cd9a7a-bb01-43de-9055-deccbc8529b2-kube-api-access-t2th5\") pod \"image-pruner-29331360-494rn\" (UID: \"84cd9a7a-bb01-43de-9055-deccbc8529b2\") " pod="openshift-image-registry/image-pruner-29331360-494rn" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.005009 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nrbdr\" (UniqueName: \"kubernetes.io/projected/bb075d42-7ef0-4ad4-b817-bdd611f0d7ed-kube-api-access-nrbdr\") pod \"apiserver-7bbb656c7d-rjwhx\" (UID: \"bb075d42-7ef0-4ad4-b817-bdd611f0d7ed\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rjwhx" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.064340 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bcxxq\" (UniqueName: \"kubernetes.io/projected/86941755-9852-4757-82b6-c4aa2c3ffa43-kube-api-access-bcxxq\") pod \"openshift-apiserver-operator-796bbdcf4f-p9jtz\" (UID: \"86941755-9852-4757-82b6-c4aa2c3ffa43\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-p9jtz" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.183985 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.197616 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-tmj7c"] Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.204188 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.222751 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.245923 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.248803 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-jjhb5"] Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.264102 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.285478 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.288709 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-tlvk7"] Oct 08 00:08:42 crc kubenswrapper[4967]: W1008 00:08:42.301142 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd74a2ad0_73e6_4f0d_a804_b7ede2cf8835.slice/crio-011205476e92f6866751843db10415bb8fb22a084933e0adfc85f99053dd1198 WatchSource:0}: Error finding container 011205476e92f6866751843db10415bb8fb22a084933e0adfc85f99053dd1198: Status 404 returned error can't find the container with id 011205476e92f6866751843db10415bb8fb22a084933e0adfc85f99053dd1198 Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.318570 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-tdqfv"] Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.322260 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.330279 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-krqks"] Oct 08 00:08:42 crc kubenswrapper[4967]: W1008 00:08:42.331214 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfa677745_a987_479c_b60a_2220f1bd58d1.slice/crio-551e116fdea35ec6dd300bbbf8a10d7a7bcc715f4c260c8bac3deb26e236c504 WatchSource:0}: Error finding container 551e116fdea35ec6dd300bbbf8a10d7a7bcc715f4c260c8bac3deb26e236c504: Status 404 returned error can't find the container with id 551e116fdea35ec6dd300bbbf8a10d7a7bcc715f4c260c8bac3deb26e236c504 Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.342535 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Oct 08 00:08:42 crc kubenswrapper[4967]: W1008 00:08:42.343100 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfccf1013_b838_4303_8da2_4861e234540b.slice/crio-db5c57086329a1b8df15856f2e0ae83c164861f24d72d8853031ad4e24fbba4d WatchSource:0}: Error finding container db5c57086329a1b8df15856f2e0ae83c164861f24d72d8853031ad4e24fbba4d: Status 404 returned error can't find the container with id db5c57086329a1b8df15856f2e0ae83c164861f24d72d8853031ad4e24fbba4d Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.362747 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.383108 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.388100 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/1f06a5d1-8ff8-4d73-8939-0fac4252ea9c-registry-tls\") pod \"image-registry-697d97f7c8-8d8x5\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.388140 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/9fa40840-4cd4-41ea-9c97-8a76e3001eee-srv-cert\") pod \"olm-operator-6b444d44fb-4kzvf\" (UID: \"9fa40840-4cd4-41ea-9c97-8a76e3001eee\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4kzvf" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.388173 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/aa6ab9dd-70bc-4834-a3be-f934e73c15c5-mountpoint-dir\") pod \"csi-hostpathplugin-t65kf\" (UID: \"aa6ab9dd-70bc-4834-a3be-f934e73c15c5\") " pod="hostpath-provisioner/csi-hostpathplugin-t65kf" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.388473 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4481eaf4-ecf4-46d3-be86-f028e4e33970-trusted-ca\") pod \"ingress-operator-5b745b69d9-wp2tt\" (UID: \"4481eaf4-ecf4-46d3-be86-f028e4e33970\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wp2tt" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.388866 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/aa6ab9dd-70bc-4834-a3be-f934e73c15c5-socket-dir\") pod \"csi-hostpathplugin-t65kf\" (UID: \"aa6ab9dd-70bc-4834-a3be-f934e73c15c5\") " pod="hostpath-provisioner/csi-hostpathplugin-t65kf" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.389098 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/9fa40840-4cd4-41ea-9c97-8a76e3001eee-profile-collector-cert\") pod \"olm-operator-6b444d44fb-4kzvf\" (UID: \"9fa40840-4cd4-41ea-9c97-8a76e3001eee\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4kzvf" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.389169 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e47422e6-8d98-4115-b59d-615218bfd4ae-config\") pod \"service-ca-operator-777779d784-kqmk8\" (UID: \"e47422e6-8d98-4115-b59d-615218bfd4ae\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-kqmk8" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.389201 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vtx28\" (UniqueName: \"kubernetes.io/projected/e47422e6-8d98-4115-b59d-615218bfd4ae-kube-api-access-vtx28\") pod \"service-ca-operator-777779d784-kqmk8\" (UID: \"e47422e6-8d98-4115-b59d-615218bfd4ae\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-kqmk8" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.389592 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9jvkr\" (UniqueName: \"kubernetes.io/projected/9876428c-dc70-42b1-886b-2c8dfff20039-kube-api-access-9jvkr\") pod \"migrator-59844c95c7-zkgkc\" (UID: \"9876428c-dc70-42b1-886b-2c8dfff20039\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-zkgkc" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.390149 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2302f69-8556-43c8-9451-145211206874-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-n69qz\" (UID: \"c2302f69-8556-43c8-9451-145211206874\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n69qz" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.390183 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b2693055-f9b3-4539-ac6d-2151aafe6c2e-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-25pkg\" (UID: \"b2693055-f9b3-4539-ac6d-2151aafe6c2e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-25pkg" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.390238 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d8rcj\" (UniqueName: \"kubernetes.io/projected/ee7dadb3-3ab7-4790-9b67-7f7a5bf7604f-kube-api-access-d8rcj\") pod \"router-default-5444994796-p9kxf\" (UID: \"ee7dadb3-3ab7-4790-9b67-7f7a5bf7604f\") " pod="openshift-ingress/router-default-5444994796-p9kxf" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.390548 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/1f06a5d1-8ff8-4d73-8939-0fac4252ea9c-ca-trust-extracted\") pod \"image-registry-697d97f7c8-8d8x5\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.390706 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/2c9d9d1d-3e68-47b5-b649-f193653c1922-signing-key\") pod \"service-ca-9c57cc56f-ghwrl\" (UID: \"2c9d9d1d-3e68-47b5-b649-f193653c1922\") " pod="openshift-service-ca/service-ca-9c57cc56f-ghwrl" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.390746 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vh779\" (UniqueName: \"kubernetes.io/projected/64202fa4-cb80-4028-aa03-d53231f2d8f4-kube-api-access-vh779\") pod \"ingress-canary-hpx62\" (UID: \"64202fa4-cb80-4028-aa03-d53231f2d8f4\") " pod="openshift-ingress-canary/ingress-canary-hpx62" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.390845 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w22ps\" (UniqueName: \"kubernetes.io/projected/f79e3965-8fdc-48ba-be6c-2866348f2458-kube-api-access-w22ps\") pod \"marketplace-operator-79b997595-npdfq\" (UID: \"f79e3965-8fdc-48ba-be6c-2866348f2458\") " pod="openshift-marketplace/marketplace-operator-79b997595-npdfq" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.390975 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/1f06a5d1-8ff8-4d73-8939-0fac4252ea9c-installation-pull-secrets\") pod \"image-registry-697d97f7c8-8d8x5\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.391009 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1f06a5d1-8ff8-4d73-8939-0fac4252ea9c-bound-sa-token\") pod \"image-registry-697d97f7c8-8d8x5\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.391041 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b2693055-f9b3-4539-ac6d-2151aafe6c2e-proxy-tls\") pod \"machine-config-controller-84d6567774-25pkg\" (UID: \"b2693055-f9b3-4539-ac6d-2151aafe6c2e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-25pkg" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.391067 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/aa6ab9dd-70bc-4834-a3be-f934e73c15c5-plugins-dir\") pod \"csi-hostpathplugin-t65kf\" (UID: \"aa6ab9dd-70bc-4834-a3be-f934e73c15c5\") " pod="hostpath-provisioner/csi-hostpathplugin-t65kf" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.391242 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vpcpf\" (UniqueName: \"kubernetes.io/projected/9fa40840-4cd4-41ea-9c97-8a76e3001eee-kube-api-access-vpcpf\") pod \"olm-operator-6b444d44fb-4kzvf\" (UID: \"9fa40840-4cd4-41ea-9c97-8a76e3001eee\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4kzvf" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.391340 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/4481eaf4-ecf4-46d3-be86-f028e4e33970-metrics-tls\") pod \"ingress-operator-5b745b69d9-wp2tt\" (UID: \"4481eaf4-ecf4-46d3-be86-f028e4e33970\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wp2tt" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.391396 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/aa6ab9dd-70bc-4834-a3be-f934e73c15c5-registration-dir\") pod \"csi-hostpathplugin-t65kf\" (UID: \"aa6ab9dd-70bc-4834-a3be-f934e73c15c5\") " pod="hostpath-provisioner/csi-hostpathplugin-t65kf" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.391763 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c7b7dad0-eb2c-4c7b-ac09-48d58ed7b936-auth-proxy-config\") pod \"machine-config-operator-74547568cd-2ghtr\" (UID: \"c7b7dad0-eb2c-4c7b-ac09-48d58ed7b936\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2ghtr" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.391933 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7e86e7c5-1431-4606-bf22-7bc05205e8e8-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-pjlzq\" (UID: \"7e86e7c5-1431-4606-bf22-7bc05205e8e8\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-pjlzq" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.391975 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ee7dadb3-3ab7-4790-9b67-7f7a5bf7604f-metrics-certs\") pod \"router-default-5444994796-p9kxf\" (UID: \"ee7dadb3-3ab7-4790-9b67-7f7a5bf7604f\") " pod="openshift-ingress/router-default-5444994796-p9kxf" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.393054 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bb260fdf-16d6-4771-b7ae-38a669ff2393-secret-volume\") pod \"collect-profiles-29331360-t4jlf\" (UID: \"bb260fdf-16d6-4771-b7ae-38a669ff2393\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331360-t4jlf" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.393137 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/8e5ce93e-573e-456c-a08c-4737fb3c7a66-console-oauth-config\") pod \"console-f9d7485db-vxdxg\" (UID: \"8e5ce93e-573e-456c-a08c-4737fb3c7a66\") " pod="openshift-console/console-f9d7485db-vxdxg" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.393199 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/8e5ce93e-573e-456c-a08c-4737fb3c7a66-console-serving-cert\") pod \"console-f9d7485db-vxdxg\" (UID: \"8e5ce93e-573e-456c-a08c-4737fb3c7a66\") " pod="openshift-console/console-f9d7485db-vxdxg" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.393245 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/ee7dadb3-3ab7-4790-9b67-7f7a5bf7604f-stats-auth\") pod \"router-default-5444994796-p9kxf\" (UID: \"ee7dadb3-3ab7-4790-9b67-7f7a5bf7604f\") " pod="openshift-ingress/router-default-5444994796-p9kxf" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.393296 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/cd33606b-2d0f-4f4b-bc86-b85ad4f56f83-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-lpxcq\" (UID: \"cd33606b-2d0f-4f4b-bc86-b85ad4f56f83\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-lpxcq" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.393584 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/1f06a5d1-8ff8-4d73-8939-0fac4252ea9c-registry-certificates\") pod \"image-registry-697d97f7c8-8d8x5\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.393686 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/8e5ce93e-573e-456c-a08c-4737fb3c7a66-oauth-serving-cert\") pod \"console-f9d7485db-vxdxg\" (UID: \"8e5ce93e-573e-456c-a08c-4737fb3c7a66\") " pod="openshift-console/console-f9d7485db-vxdxg" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.393821 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e873b76e-e526-4a38-a2d2-9b23ac147852-etcd-client\") pod \"etcd-operator-b45778765-fqmxp\" (UID: \"e873b76e-e526-4a38-a2d2-9b23ac147852\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fqmxp" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.393860 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hbmqm\" (UniqueName: \"kubernetes.io/projected/e0d38c54-b3a4-46b1-af38-0f8c7d3dbf5f-kube-api-access-hbmqm\") pod \"package-server-manager-789f6589d5-b5glr\" (UID: \"e0d38c54-b3a4-46b1-af38-0f8c7d3dbf5f\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b5glr" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.393983 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7e86e7c5-1431-4606-bf22-7bc05205e8e8-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-pjlzq\" (UID: \"7e86e7c5-1431-4606-bf22-7bc05205e8e8\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-pjlzq" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.394169 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vpr4c\" (UniqueName: \"kubernetes.io/projected/c7b7dad0-eb2c-4c7b-ac09-48d58ed7b936-kube-api-access-vpr4c\") pod \"machine-config-operator-74547568cd-2ghtr\" (UID: \"c7b7dad0-eb2c-4c7b-ac09-48d58ed7b936\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2ghtr" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.394219 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4481eaf4-ecf4-46d3-be86-f028e4e33970-bound-sa-token\") pod \"ingress-operator-5b745b69d9-wp2tt\" (UID: \"4481eaf4-ecf4-46d3-be86-f028e4e33970\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wp2tt" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.394268 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/2c9d9d1d-3e68-47b5-b649-f193653c1922-signing-cabundle\") pod \"service-ca-9c57cc56f-ghwrl\" (UID: \"2c9d9d1d-3e68-47b5-b649-f193653c1922\") " pod="openshift-service-ca/service-ca-9c57cc56f-ghwrl" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.394349 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/e873b76e-e526-4a38-a2d2-9b23ac147852-etcd-ca\") pod \"etcd-operator-b45778765-fqmxp\" (UID: \"e873b76e-e526-4a38-a2d2-9b23ac147852\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fqmxp" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.394647 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8e5ce93e-573e-456c-a08c-4737fb3c7a66-trusted-ca-bundle\") pod \"console-f9d7485db-vxdxg\" (UID: \"8e5ce93e-573e-456c-a08c-4737fb3c7a66\") " pod="openshift-console/console-f9d7485db-vxdxg" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.394724 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/e0d38c54-b3a4-46b1-af38-0f8c7d3dbf5f-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-b5glr\" (UID: \"e0d38c54-b3a4-46b1-af38-0f8c7d3dbf5f\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b5glr" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.395711 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1f06a5d1-8ff8-4d73-8939-0fac4252ea9c-trusted-ca\") pod \"image-registry-697d97f7c8-8d8x5\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.395748 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/e873b76e-e526-4a38-a2d2-9b23ac147852-etcd-service-ca\") pod \"etcd-operator-b45778765-fqmxp\" (UID: \"e873b76e-e526-4a38-a2d2-9b23ac147852\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fqmxp" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.395789 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hkk8l\" (UniqueName: \"kubernetes.io/projected/cd33606b-2d0f-4f4b-bc86-b85ad4f56f83-kube-api-access-hkk8l\") pod \"control-plane-machine-set-operator-78cbb6b69f-lpxcq\" (UID: \"cd33606b-2d0f-4f4b-bc86-b85ad4f56f83\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-lpxcq" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.395832 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/c7b7dad0-eb2c-4c7b-ac09-48d58ed7b936-images\") pod \"machine-config-operator-74547568cd-2ghtr\" (UID: \"c7b7dad0-eb2c-4c7b-ac09-48d58ed7b936\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2ghtr" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.395907 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vcvj4\" (UniqueName: \"kubernetes.io/projected/4481eaf4-ecf4-46d3-be86-f028e4e33970-kube-api-access-vcvj4\") pod \"ingress-operator-5b745b69d9-wp2tt\" (UID: \"4481eaf4-ecf4-46d3-be86-f028e4e33970\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wp2tt" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.395946 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/8e5ce93e-573e-456c-a08c-4737fb3c7a66-console-config\") pod \"console-f9d7485db-vxdxg\" (UID: \"8e5ce93e-573e-456c-a08c-4737fb3c7a66\") " pod="openshift-console/console-f9d7485db-vxdxg" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.395969 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/ee7dadb3-3ab7-4790-9b67-7f7a5bf7604f-default-certificate\") pod \"router-default-5444994796-p9kxf\" (UID: \"ee7dadb3-3ab7-4790-9b67-7f7a5bf7604f\") " pod="openshift-ingress/router-default-5444994796-p9kxf" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.395996 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/c7b7dad0-eb2c-4c7b-ac09-48d58ed7b936-proxy-tls\") pod \"machine-config-operator-74547568cd-2ghtr\" (UID: \"c7b7dad0-eb2c-4c7b-ac09-48d58ed7b936\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2ghtr" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.396018 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/aa6ab9dd-70bc-4834-a3be-f934e73c15c5-csi-data-dir\") pod \"csi-hostpathplugin-t65kf\" (UID: \"aa6ab9dd-70bc-4834-a3be-f934e73c15c5\") " pod="hostpath-provisioner/csi-hostpathplugin-t65kf" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.396051 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8d8x5\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.396078 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/553037a9-2f32-4eee-8244-caa084e58566-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-86lvz\" (UID: \"553037a9-2f32-4eee-8244-caa084e58566\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-86lvz" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.396102 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e873b76e-e526-4a38-a2d2-9b23ac147852-serving-cert\") pod \"etcd-operator-b45778765-fqmxp\" (UID: \"e873b76e-e526-4a38-a2d2-9b23ac147852\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fqmxp" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.396126 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/64202fa4-cb80-4028-aa03-d53231f2d8f4-cert\") pod \"ingress-canary-hpx62\" (UID: \"64202fa4-cb80-4028-aa03-d53231f2d8f4\") " pod="openshift-ingress-canary/ingress-canary-hpx62" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.396165 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e47422e6-8d98-4115-b59d-615218bfd4ae-serving-cert\") pod \"service-ca-operator-777779d784-kqmk8\" (UID: \"e47422e6-8d98-4115-b59d-615218bfd4ae\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-kqmk8" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.396193 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zt77j\" (UniqueName: \"kubernetes.io/projected/553037a9-2f32-4eee-8244-caa084e58566-kube-api-access-zt77j\") pod \"multus-admission-controller-857f4d67dd-86lvz\" (UID: \"553037a9-2f32-4eee-8244-caa084e58566\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-86lvz" Oct 08 00:08:42 crc kubenswrapper[4967]: E1008 00:08:42.396553 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 00:08:42.896535831 +0000 UTC m=+121.547070480 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8d8x5" (UID: "1f06a5d1-8ff8-4d73-8939-0fac4252ea9c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.396860 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9htfq\" (UniqueName: \"kubernetes.io/projected/7e86e7c5-1431-4606-bf22-7bc05205e8e8-kube-api-access-9htfq\") pod \"cluster-image-registry-operator-dc59b4c8b-pjlzq\" (UID: \"7e86e7c5-1431-4606-bf22-7bc05205e8e8\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-pjlzq" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.397602 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c2302f69-8556-43c8-9451-145211206874-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-n69qz\" (UID: \"c2302f69-8556-43c8-9451-145211206874\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n69qz" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.397643 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/7e86e7c5-1431-4606-bf22-7bc05205e8e8-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-pjlzq\" (UID: \"7e86e7c5-1431-4606-bf22-7bc05205e8e8\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-pjlzq" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.397720 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8e5ce93e-573e-456c-a08c-4737fb3c7a66-service-ca\") pod \"console-f9d7485db-vxdxg\" (UID: \"8e5ce93e-573e-456c-a08c-4737fb3c7a66\") " pod="openshift-console/console-f9d7485db-vxdxg" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.397831 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ee7dadb3-3ab7-4790-9b67-7f7a5bf7604f-service-ca-bundle\") pod \"router-default-5444994796-p9kxf\" (UID: \"ee7dadb3-3ab7-4790-9b67-7f7a5bf7604f\") " pod="openshift-ingress/router-default-5444994796-p9kxf" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.397859 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f79e3965-8fdc-48ba-be6c-2866348f2458-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-npdfq\" (UID: \"f79e3965-8fdc-48ba-be6c-2866348f2458\") " pod="openshift-marketplace/marketplace-operator-79b997595-npdfq" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.397892 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bb260fdf-16d6-4771-b7ae-38a669ff2393-config-volume\") pod \"collect-profiles-29331360-t4jlf\" (UID: \"bb260fdf-16d6-4771-b7ae-38a669ff2393\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331360-t4jlf" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.397913 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c2302f69-8556-43c8-9451-145211206874-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-n69qz\" (UID: \"c2302f69-8556-43c8-9451-145211206874\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n69qz" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.398121 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hkw6t\" (UniqueName: \"kubernetes.io/projected/2c9d9d1d-3e68-47b5-b649-f193653c1922-kube-api-access-hkw6t\") pod \"service-ca-9c57cc56f-ghwrl\" (UID: \"2c9d9d1d-3e68-47b5-b649-f193653c1922\") " pod="openshift-service-ca/service-ca-9c57cc56f-ghwrl" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.398165 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-br729\" (UniqueName: \"kubernetes.io/projected/aa6ab9dd-70bc-4834-a3be-f934e73c15c5-kube-api-access-br729\") pod \"csi-hostpathplugin-t65kf\" (UID: \"aa6ab9dd-70bc-4834-a3be-f934e73c15c5\") " pod="hostpath-provisioner/csi-hostpathplugin-t65kf" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.398259 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lbttt\" (UniqueName: \"kubernetes.io/projected/8e5ce93e-573e-456c-a08c-4737fb3c7a66-kube-api-access-lbttt\") pod \"console-f9d7485db-vxdxg\" (UID: \"8e5ce93e-573e-456c-a08c-4737fb3c7a66\") " pod="openshift-console/console-f9d7485db-vxdxg" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.398612 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6lf6\" (UniqueName: \"kubernetes.io/projected/1f06a5d1-8ff8-4d73-8939-0fac4252ea9c-kube-api-access-b6lf6\") pod \"image-registry-697d97f7c8-8d8x5\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.398684 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f79e3965-8fdc-48ba-be6c-2866348f2458-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-npdfq\" (UID: \"f79e3965-8fdc-48ba-be6c-2866348f2458\") " pod="openshift-marketplace/marketplace-operator-79b997595-npdfq" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.398712 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e873b76e-e526-4a38-a2d2-9b23ac147852-config\") pod \"etcd-operator-b45778765-fqmxp\" (UID: \"e873b76e-e526-4a38-a2d2-9b23ac147852\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fqmxp" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.398731 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-64h5g\" (UniqueName: \"kubernetes.io/projected/b2693055-f9b3-4539-ac6d-2151aafe6c2e-kube-api-access-64h5g\") pod \"machine-config-controller-84d6567774-25pkg\" (UID: \"b2693055-f9b3-4539-ac6d-2151aafe6c2e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-25pkg" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.400570 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c7plb\" (UniqueName: \"kubernetes.io/projected/e873b76e-e526-4a38-a2d2-9b23ac147852-kube-api-access-c7plb\") pod \"etcd-operator-b45778765-fqmxp\" (UID: \"e873b76e-e526-4a38-a2d2-9b23ac147852\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fqmxp" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.400699 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-48bwp\" (UniqueName: \"kubernetes.io/projected/bb260fdf-16d6-4771-b7ae-38a669ff2393-kube-api-access-48bwp\") pod \"collect-profiles-29331360-t4jlf\" (UID: \"bb260fdf-16d6-4771-b7ae-38a669ff2393\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331360-t4jlf" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.402712 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.404741 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-ljjfc" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.422750 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.429597 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-ck4pd" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.444475 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.457286 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f3ea161f-2b22-466c-b49b-9f60e9e3437a-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-ltttx\" (UID: \"f3ea161f-2b22-466c-b49b-9f60e9e3437a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ltttx" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.464977 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.471054 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0d5e578e-60b8-400c-9545-110600602c3c-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-tzhwk\" (UID: \"0d5e578e-60b8-400c-9545-110600602c3c\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-tzhwk" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.483429 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.483877 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-b5mzq" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.505197 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 00:08:42 crc kubenswrapper[4967]: E1008 00:08:42.505389 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 00:08:43.005369371 +0000 UTC m=+121.655904020 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.511870 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/9fa40840-4cd4-41ea-9c97-8a76e3001eee-profile-collector-cert\") pod \"olm-operator-6b444d44fb-4kzvf\" (UID: \"9fa40840-4cd4-41ea-9c97-8a76e3001eee\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4kzvf" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.511920 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e47422e6-8d98-4115-b59d-615218bfd4ae-config\") pod \"service-ca-operator-777779d784-kqmk8\" (UID: \"e47422e6-8d98-4115-b59d-615218bfd4ae\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-kqmk8" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.511944 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vtx28\" (UniqueName: \"kubernetes.io/projected/e47422e6-8d98-4115-b59d-615218bfd4ae-kube-api-access-vtx28\") pod \"service-ca-operator-777779d784-kqmk8\" (UID: \"e47422e6-8d98-4115-b59d-615218bfd4ae\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-kqmk8" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.511961 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9jvkr\" (UniqueName: \"kubernetes.io/projected/9876428c-dc70-42b1-886b-2c8dfff20039-kube-api-access-9jvkr\") pod \"migrator-59844c95c7-zkgkc\" (UID: \"9876428c-dc70-42b1-886b-2c8dfff20039\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-zkgkc" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.512005 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2302f69-8556-43c8-9451-145211206874-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-n69qz\" (UID: \"c2302f69-8556-43c8-9451-145211206874\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n69qz" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.512026 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b2693055-f9b3-4539-ac6d-2151aafe6c2e-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-25pkg\" (UID: \"b2693055-f9b3-4539-ac6d-2151aafe6c2e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-25pkg" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.512047 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d8rcj\" (UniqueName: \"kubernetes.io/projected/ee7dadb3-3ab7-4790-9b67-7f7a5bf7604f-kube-api-access-d8rcj\") pod \"router-default-5444994796-p9kxf\" (UID: \"ee7dadb3-3ab7-4790-9b67-7f7a5bf7604f\") " pod="openshift-ingress/router-default-5444994796-p9kxf" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.512095 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/1f06a5d1-8ff8-4d73-8939-0fac4252ea9c-ca-trust-extracted\") pod \"image-registry-697d97f7c8-8d8x5\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.512121 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/2c9d9d1d-3e68-47b5-b649-f193653c1922-signing-key\") pod \"service-ca-9c57cc56f-ghwrl\" (UID: \"2c9d9d1d-3e68-47b5-b649-f193653c1922\") " pod="openshift-service-ca/service-ca-9c57cc56f-ghwrl" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.512138 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vh779\" (UniqueName: \"kubernetes.io/projected/64202fa4-cb80-4028-aa03-d53231f2d8f4-kube-api-access-vh779\") pod \"ingress-canary-hpx62\" (UID: \"64202fa4-cb80-4028-aa03-d53231f2d8f4\") " pod="openshift-ingress-canary/ingress-canary-hpx62" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.512154 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w22ps\" (UniqueName: \"kubernetes.io/projected/f79e3965-8fdc-48ba-be6c-2866348f2458-kube-api-access-w22ps\") pod \"marketplace-operator-79b997595-npdfq\" (UID: \"f79e3965-8fdc-48ba-be6c-2866348f2458\") " pod="openshift-marketplace/marketplace-operator-79b997595-npdfq" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.512188 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/1f06a5d1-8ff8-4d73-8939-0fac4252ea9c-installation-pull-secrets\") pod \"image-registry-697d97f7c8-8d8x5\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.512205 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1f06a5d1-8ff8-4d73-8939-0fac4252ea9c-bound-sa-token\") pod \"image-registry-697d97f7c8-8d8x5\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.512223 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b2693055-f9b3-4539-ac6d-2151aafe6c2e-proxy-tls\") pod \"machine-config-controller-84d6567774-25pkg\" (UID: \"b2693055-f9b3-4539-ac6d-2151aafe6c2e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-25pkg" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.512241 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/aa6ab9dd-70bc-4834-a3be-f934e73c15c5-plugins-dir\") pod \"csi-hostpathplugin-t65kf\" (UID: \"aa6ab9dd-70bc-4834-a3be-f934e73c15c5\") " pod="hostpath-provisioner/csi-hostpathplugin-t65kf" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.512257 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vpcpf\" (UniqueName: \"kubernetes.io/projected/9fa40840-4cd4-41ea-9c97-8a76e3001eee-kube-api-access-vpcpf\") pod \"olm-operator-6b444d44fb-4kzvf\" (UID: \"9fa40840-4cd4-41ea-9c97-8a76e3001eee\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4kzvf" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.512275 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/4481eaf4-ecf4-46d3-be86-f028e4e33970-metrics-tls\") pod \"ingress-operator-5b745b69d9-wp2tt\" (UID: \"4481eaf4-ecf4-46d3-be86-f028e4e33970\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wp2tt" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.512289 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/aa6ab9dd-70bc-4834-a3be-f934e73c15c5-registration-dir\") pod \"csi-hostpathplugin-t65kf\" (UID: \"aa6ab9dd-70bc-4834-a3be-f934e73c15c5\") " pod="hostpath-provisioner/csi-hostpathplugin-t65kf" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.512306 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c7b7dad0-eb2c-4c7b-ac09-48d58ed7b936-auth-proxy-config\") pod \"machine-config-operator-74547568cd-2ghtr\" (UID: \"c7b7dad0-eb2c-4c7b-ac09-48d58ed7b936\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2ghtr" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.512326 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7e86e7c5-1431-4606-bf22-7bc05205e8e8-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-pjlzq\" (UID: \"7e86e7c5-1431-4606-bf22-7bc05205e8e8\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-pjlzq" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.512342 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ee7dadb3-3ab7-4790-9b67-7f7a5bf7604f-metrics-certs\") pod \"router-default-5444994796-p9kxf\" (UID: \"ee7dadb3-3ab7-4790-9b67-7f7a5bf7604f\") " pod="openshift-ingress/router-default-5444994796-p9kxf" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.512374 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bb260fdf-16d6-4771-b7ae-38a669ff2393-secret-volume\") pod \"collect-profiles-29331360-t4jlf\" (UID: \"bb260fdf-16d6-4771-b7ae-38a669ff2393\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331360-t4jlf" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.512389 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/8e5ce93e-573e-456c-a08c-4737fb3c7a66-console-oauth-config\") pod \"console-f9d7485db-vxdxg\" (UID: \"8e5ce93e-573e-456c-a08c-4737fb3c7a66\") " pod="openshift-console/console-f9d7485db-vxdxg" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.512418 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/ee7dadb3-3ab7-4790-9b67-7f7a5bf7604f-stats-auth\") pod \"router-default-5444994796-p9kxf\" (UID: \"ee7dadb3-3ab7-4790-9b67-7f7a5bf7604f\") " pod="openshift-ingress/router-default-5444994796-p9kxf" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.512433 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/8e5ce93e-573e-456c-a08c-4737fb3c7a66-console-serving-cert\") pod \"console-f9d7485db-vxdxg\" (UID: \"8e5ce93e-573e-456c-a08c-4737fb3c7a66\") " pod="openshift-console/console-f9d7485db-vxdxg" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.512450 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/cd33606b-2d0f-4f4b-bc86-b85ad4f56f83-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-lpxcq\" (UID: \"cd33606b-2d0f-4f4b-bc86-b85ad4f56f83\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-lpxcq" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.512466 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/1f06a5d1-8ff8-4d73-8939-0fac4252ea9c-registry-certificates\") pod \"image-registry-697d97f7c8-8d8x5\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.512482 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/8e5ce93e-573e-456c-a08c-4737fb3c7a66-oauth-serving-cert\") pod \"console-f9d7485db-vxdxg\" (UID: \"8e5ce93e-573e-456c-a08c-4737fb3c7a66\") " pod="openshift-console/console-f9d7485db-vxdxg" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.512510 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e873b76e-e526-4a38-a2d2-9b23ac147852-etcd-client\") pod \"etcd-operator-b45778765-fqmxp\" (UID: \"e873b76e-e526-4a38-a2d2-9b23ac147852\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fqmxp" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.512550 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hbmqm\" (UniqueName: \"kubernetes.io/projected/e0d38c54-b3a4-46b1-af38-0f8c7d3dbf5f-kube-api-access-hbmqm\") pod \"package-server-manager-789f6589d5-b5glr\" (UID: \"e0d38c54-b3a4-46b1-af38-0f8c7d3dbf5f\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b5glr" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.512571 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7e86e7c5-1431-4606-bf22-7bc05205e8e8-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-pjlzq\" (UID: \"7e86e7c5-1431-4606-bf22-7bc05205e8e8\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-pjlzq" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.512595 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vpr4c\" (UniqueName: \"kubernetes.io/projected/c7b7dad0-eb2c-4c7b-ac09-48d58ed7b936-kube-api-access-vpr4c\") pod \"machine-config-operator-74547568cd-2ghtr\" (UID: \"c7b7dad0-eb2c-4c7b-ac09-48d58ed7b936\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2ghtr" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.512613 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4481eaf4-ecf4-46d3-be86-f028e4e33970-bound-sa-token\") pod \"ingress-operator-5b745b69d9-wp2tt\" (UID: \"4481eaf4-ecf4-46d3-be86-f028e4e33970\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wp2tt" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.512628 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/2c9d9d1d-3e68-47b5-b649-f193653c1922-signing-cabundle\") pod \"service-ca-9c57cc56f-ghwrl\" (UID: \"2c9d9d1d-3e68-47b5-b649-f193653c1922\") " pod="openshift-service-ca/service-ca-9c57cc56f-ghwrl" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.512645 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/e873b76e-e526-4a38-a2d2-9b23ac147852-etcd-ca\") pod \"etcd-operator-b45778765-fqmxp\" (UID: \"e873b76e-e526-4a38-a2d2-9b23ac147852\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fqmxp" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.512695 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/bb6ecea9-8eb0-4ff6-a898-5cb24d813314-certs\") pod \"machine-config-server-h786z\" (UID: \"bb6ecea9-8eb0-4ff6-a898-5cb24d813314\") " pod="openshift-machine-config-operator/machine-config-server-h786z" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.512715 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8e5ce93e-573e-456c-a08c-4737fb3c7a66-trusted-ca-bundle\") pod \"console-f9d7485db-vxdxg\" (UID: \"8e5ce93e-573e-456c-a08c-4737fb3c7a66\") " pod="openshift-console/console-f9d7485db-vxdxg" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.512738 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/e0d38c54-b3a4-46b1-af38-0f8c7d3dbf5f-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-b5glr\" (UID: \"e0d38c54-b3a4-46b1-af38-0f8c7d3dbf5f\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b5glr" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.512756 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gnznl\" (UniqueName: \"kubernetes.io/projected/25199d70-97f2-4513-a2e4-301c9ce3f85a-kube-api-access-gnznl\") pod \"dns-default-fjz5z\" (UID: \"25199d70-97f2-4513-a2e4-301c9ce3f85a\") " pod="openshift-dns/dns-default-fjz5z" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.512779 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1f06a5d1-8ff8-4d73-8939-0fac4252ea9c-trusted-ca\") pod \"image-registry-697d97f7c8-8d8x5\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.512793 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/e873b76e-e526-4a38-a2d2-9b23ac147852-etcd-service-ca\") pod \"etcd-operator-b45778765-fqmxp\" (UID: \"e873b76e-e526-4a38-a2d2-9b23ac147852\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fqmxp" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.512807 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/25199d70-97f2-4513-a2e4-301c9ce3f85a-config-volume\") pod \"dns-default-fjz5z\" (UID: \"25199d70-97f2-4513-a2e4-301c9ce3f85a\") " pod="openshift-dns/dns-default-fjz5z" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.512825 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hkk8l\" (UniqueName: \"kubernetes.io/projected/cd33606b-2d0f-4f4b-bc86-b85ad4f56f83-kube-api-access-hkk8l\") pod \"control-plane-machine-set-operator-78cbb6b69f-lpxcq\" (UID: \"cd33606b-2d0f-4f4b-bc86-b85ad4f56f83\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-lpxcq" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.512851 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/c7b7dad0-eb2c-4c7b-ac09-48d58ed7b936-images\") pod \"machine-config-operator-74547568cd-2ghtr\" (UID: \"c7b7dad0-eb2c-4c7b-ac09-48d58ed7b936\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2ghtr" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.512871 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vcvj4\" (UniqueName: \"kubernetes.io/projected/4481eaf4-ecf4-46d3-be86-f028e4e33970-kube-api-access-vcvj4\") pod \"ingress-operator-5b745b69d9-wp2tt\" (UID: \"4481eaf4-ecf4-46d3-be86-f028e4e33970\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wp2tt" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.512898 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/8e5ce93e-573e-456c-a08c-4737fb3c7a66-console-config\") pod \"console-f9d7485db-vxdxg\" (UID: \"8e5ce93e-573e-456c-a08c-4737fb3c7a66\") " pod="openshift-console/console-f9d7485db-vxdxg" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.512917 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/ee7dadb3-3ab7-4790-9b67-7f7a5bf7604f-default-certificate\") pod \"router-default-5444994796-p9kxf\" (UID: \"ee7dadb3-3ab7-4790-9b67-7f7a5bf7604f\") " pod="openshift-ingress/router-default-5444994796-p9kxf" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.512932 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/c7b7dad0-eb2c-4c7b-ac09-48d58ed7b936-proxy-tls\") pod \"machine-config-operator-74547568cd-2ghtr\" (UID: \"c7b7dad0-eb2c-4c7b-ac09-48d58ed7b936\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2ghtr" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.512947 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/aa6ab9dd-70bc-4834-a3be-f934e73c15c5-csi-data-dir\") pod \"csi-hostpathplugin-t65kf\" (UID: \"aa6ab9dd-70bc-4834-a3be-f934e73c15c5\") " pod="hostpath-provisioner/csi-hostpathplugin-t65kf" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.512964 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e873b76e-e526-4a38-a2d2-9b23ac147852-serving-cert\") pod \"etcd-operator-b45778765-fqmxp\" (UID: \"e873b76e-e526-4a38-a2d2-9b23ac147852\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fqmxp" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.512977 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/64202fa4-cb80-4028-aa03-d53231f2d8f4-cert\") pod \"ingress-canary-hpx62\" (UID: \"64202fa4-cb80-4028-aa03-d53231f2d8f4\") " pod="openshift-ingress-canary/ingress-canary-hpx62" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.512995 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8d8x5\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.513010 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/553037a9-2f32-4eee-8244-caa084e58566-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-86lvz\" (UID: \"553037a9-2f32-4eee-8244-caa084e58566\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-86lvz" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.513027 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e47422e6-8d98-4115-b59d-615218bfd4ae-serving-cert\") pod \"service-ca-operator-777779d784-kqmk8\" (UID: \"e47422e6-8d98-4115-b59d-615218bfd4ae\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-kqmk8" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.513163 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zt77j\" (UniqueName: \"kubernetes.io/projected/553037a9-2f32-4eee-8244-caa084e58566-kube-api-access-zt77j\") pod \"multus-admission-controller-857f4d67dd-86lvz\" (UID: \"553037a9-2f32-4eee-8244-caa084e58566\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-86lvz" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.513188 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9htfq\" (UniqueName: \"kubernetes.io/projected/7e86e7c5-1431-4606-bf22-7bc05205e8e8-kube-api-access-9htfq\") pod \"cluster-image-registry-operator-dc59b4c8b-pjlzq\" (UID: \"7e86e7c5-1431-4606-bf22-7bc05205e8e8\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-pjlzq" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.513226 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c2302f69-8556-43c8-9451-145211206874-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-n69qz\" (UID: \"c2302f69-8556-43c8-9451-145211206874\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n69qz" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.513247 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/25199d70-97f2-4513-a2e4-301c9ce3f85a-metrics-tls\") pod \"dns-default-fjz5z\" (UID: \"25199d70-97f2-4513-a2e4-301c9ce3f85a\") " pod="openshift-dns/dns-default-fjz5z" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.513277 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/7e86e7c5-1431-4606-bf22-7bc05205e8e8-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-pjlzq\" (UID: \"7e86e7c5-1431-4606-bf22-7bc05205e8e8\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-pjlzq" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.513293 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8e5ce93e-573e-456c-a08c-4737fb3c7a66-service-ca\") pod \"console-f9d7485db-vxdxg\" (UID: \"8e5ce93e-573e-456c-a08c-4737fb3c7a66\") " pod="openshift-console/console-f9d7485db-vxdxg" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.513324 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ee7dadb3-3ab7-4790-9b67-7f7a5bf7604f-service-ca-bundle\") pod \"router-default-5444994796-p9kxf\" (UID: \"ee7dadb3-3ab7-4790-9b67-7f7a5bf7604f\") " pod="openshift-ingress/router-default-5444994796-p9kxf" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.513345 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f79e3965-8fdc-48ba-be6c-2866348f2458-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-npdfq\" (UID: \"f79e3965-8fdc-48ba-be6c-2866348f2458\") " pod="openshift-marketplace/marketplace-operator-79b997595-npdfq" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.513367 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bb260fdf-16d6-4771-b7ae-38a669ff2393-config-volume\") pod \"collect-profiles-29331360-t4jlf\" (UID: \"bb260fdf-16d6-4771-b7ae-38a669ff2393\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331360-t4jlf" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.513383 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c2302f69-8556-43c8-9451-145211206874-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-n69qz\" (UID: \"c2302f69-8556-43c8-9451-145211206874\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n69qz" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.513416 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hkw6t\" (UniqueName: \"kubernetes.io/projected/2c9d9d1d-3e68-47b5-b649-f193653c1922-kube-api-access-hkw6t\") pod \"service-ca-9c57cc56f-ghwrl\" (UID: \"2c9d9d1d-3e68-47b5-b649-f193653c1922\") " pod="openshift-service-ca/service-ca-9c57cc56f-ghwrl" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.513435 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-br729\" (UniqueName: \"kubernetes.io/projected/aa6ab9dd-70bc-4834-a3be-f934e73c15c5-kube-api-access-br729\") pod \"csi-hostpathplugin-t65kf\" (UID: \"aa6ab9dd-70bc-4834-a3be-f934e73c15c5\") " pod="hostpath-provisioner/csi-hostpathplugin-t65kf" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.513462 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lbttt\" (UniqueName: \"kubernetes.io/projected/8e5ce93e-573e-456c-a08c-4737fb3c7a66-kube-api-access-lbttt\") pod \"console-f9d7485db-vxdxg\" (UID: \"8e5ce93e-573e-456c-a08c-4737fb3c7a66\") " pod="openshift-console/console-f9d7485db-vxdxg" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.513483 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6lf6\" (UniqueName: \"kubernetes.io/projected/1f06a5d1-8ff8-4d73-8939-0fac4252ea9c-kube-api-access-b6lf6\") pod \"image-registry-697d97f7c8-8d8x5\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.513499 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/bb6ecea9-8eb0-4ff6-a898-5cb24d813314-node-bootstrap-token\") pod \"machine-config-server-h786z\" (UID: \"bb6ecea9-8eb0-4ff6-a898-5cb24d813314\") " pod="openshift-machine-config-operator/machine-config-server-h786z" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.513516 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f79e3965-8fdc-48ba-be6c-2866348f2458-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-npdfq\" (UID: \"f79e3965-8fdc-48ba-be6c-2866348f2458\") " pod="openshift-marketplace/marketplace-operator-79b997595-npdfq" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.513533 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e873b76e-e526-4a38-a2d2-9b23ac147852-config\") pod \"etcd-operator-b45778765-fqmxp\" (UID: \"e873b76e-e526-4a38-a2d2-9b23ac147852\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fqmxp" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.513549 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-64h5g\" (UniqueName: \"kubernetes.io/projected/b2693055-f9b3-4539-ac6d-2151aafe6c2e-kube-api-access-64h5g\") pod \"machine-config-controller-84d6567774-25pkg\" (UID: \"b2693055-f9b3-4539-ac6d-2151aafe6c2e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-25pkg" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.513569 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c7plb\" (UniqueName: \"kubernetes.io/projected/e873b76e-e526-4a38-a2d2-9b23ac147852-kube-api-access-c7plb\") pod \"etcd-operator-b45778765-fqmxp\" (UID: \"e873b76e-e526-4a38-a2d2-9b23ac147852\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fqmxp" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.513591 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jx6nz\" (UniqueName: \"kubernetes.io/projected/bb6ecea9-8eb0-4ff6-a898-5cb24d813314-kube-api-access-jx6nz\") pod \"machine-config-server-h786z\" (UID: \"bb6ecea9-8eb0-4ff6-a898-5cb24d813314\") " pod="openshift-machine-config-operator/machine-config-server-h786z" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.513612 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-48bwp\" (UniqueName: \"kubernetes.io/projected/bb260fdf-16d6-4771-b7ae-38a669ff2393-kube-api-access-48bwp\") pod \"collect-profiles-29331360-t4jlf\" (UID: \"bb260fdf-16d6-4771-b7ae-38a669ff2393\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331360-t4jlf" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.513727 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/1f06a5d1-8ff8-4d73-8939-0fac4252ea9c-registry-tls\") pod \"image-registry-697d97f7c8-8d8x5\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.514074 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/9fa40840-4cd4-41ea-9c97-8a76e3001eee-srv-cert\") pod \"olm-operator-6b444d44fb-4kzvf\" (UID: \"9fa40840-4cd4-41ea-9c97-8a76e3001eee\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4kzvf" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.514098 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/aa6ab9dd-70bc-4834-a3be-f934e73c15c5-mountpoint-dir\") pod \"csi-hostpathplugin-t65kf\" (UID: \"aa6ab9dd-70bc-4834-a3be-f934e73c15c5\") " pod="hostpath-provisioner/csi-hostpathplugin-t65kf" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.514118 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/aa6ab9dd-70bc-4834-a3be-f934e73c15c5-socket-dir\") pod \"csi-hostpathplugin-t65kf\" (UID: \"aa6ab9dd-70bc-4834-a3be-f934e73c15c5\") " pod="hostpath-provisioner/csi-hostpathplugin-t65kf" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.514136 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4481eaf4-ecf4-46d3-be86-f028e4e33970-trusted-ca\") pod \"ingress-operator-5b745b69d9-wp2tt\" (UID: \"4481eaf4-ecf4-46d3-be86-f028e4e33970\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wp2tt" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.513759 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/aa6ab9dd-70bc-4834-a3be-f934e73c15c5-plugins-dir\") pod \"csi-hostpathplugin-t65kf\" (UID: \"aa6ab9dd-70bc-4834-a3be-f934e73c15c5\") " pod="hostpath-provisioner/csi-hostpathplugin-t65kf" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.514971 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/aa6ab9dd-70bc-4834-a3be-f934e73c15c5-csi-data-dir\") pod \"csi-hostpathplugin-t65kf\" (UID: \"aa6ab9dd-70bc-4834-a3be-f934e73c15c5\") " pod="hostpath-provisioner/csi-hostpathplugin-t65kf" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.513942 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/aa6ab9dd-70bc-4834-a3be-f934e73c15c5-registration-dir\") pod \"csi-hostpathplugin-t65kf\" (UID: \"aa6ab9dd-70bc-4834-a3be-f934e73c15c5\") " pod="hostpath-provisioner/csi-hostpathplugin-t65kf" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.515527 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/aa6ab9dd-70bc-4834-a3be-f934e73c15c5-mountpoint-dir\") pod \"csi-hostpathplugin-t65kf\" (UID: \"aa6ab9dd-70bc-4834-a3be-f934e73c15c5\") " pod="hostpath-provisioner/csi-hostpathplugin-t65kf" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.515623 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/aa6ab9dd-70bc-4834-a3be-f934e73c15c5-socket-dir\") pod \"csi-hostpathplugin-t65kf\" (UID: \"aa6ab9dd-70bc-4834-a3be-f934e73c15c5\") " pod="hostpath-provisioner/csi-hostpathplugin-t65kf" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.517684 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/1f06a5d1-8ff8-4d73-8939-0fac4252ea9c-registry-certificates\") pod \"image-registry-697d97f7c8-8d8x5\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:08:42 crc kubenswrapper[4967]: E1008 00:08:42.517994 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 00:08:43.017981703 +0000 UTC m=+121.668516352 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8d8x5" (UID: "1f06a5d1-8ff8-4d73-8939-0fac4252ea9c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.520189 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/1f06a5d1-8ff8-4d73-8939-0fac4252ea9c-ca-trust-extracted\") pod \"image-registry-697d97f7c8-8d8x5\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.522642 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.523496 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c7b7dad0-eb2c-4c7b-ac09-48d58ed7b936-auth-proxy-config\") pod \"machine-config-operator-74547568cd-2ghtr\" (UID: \"c7b7dad0-eb2c-4c7b-ac09-48d58ed7b936\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2ghtr" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.525064 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b2693055-f9b3-4539-ac6d-2151aafe6c2e-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-25pkg\" (UID: \"b2693055-f9b3-4539-ac6d-2151aafe6c2e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-25pkg" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.525715 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bb260fdf-16d6-4771-b7ae-38a669ff2393-secret-volume\") pod \"collect-profiles-29331360-t4jlf\" (UID: \"bb260fdf-16d6-4771-b7ae-38a669ff2393\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331360-t4jlf" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.525847 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/9fa40840-4cd4-41ea-9c97-8a76e3001eee-profile-collector-cert\") pod \"olm-operator-6b444d44fb-4kzvf\" (UID: \"9fa40840-4cd4-41ea-9c97-8a76e3001eee\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4kzvf" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.525926 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-p9jtz" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.546313 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.553663 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rjwhx" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.562291 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"pruner-dockercfg-p7bcw" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.566962 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-pruner-29331360-494rn" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.589014 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.598356 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjcsg\" (UniqueName: \"kubernetes.io/projected/0aec761a-1efe-4817-9bca-cf5961e28ac5-kube-api-access-pjcsg\") pod \"catalog-operator-68c6474976-cpnmt\" (UID: \"0aec761a-1efe-4817-9bca-cf5961e28ac5\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cpnmt" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.601930 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xtqdk\" (UniqueName: \"kubernetes.io/projected/6a6c3149-2d76-48de-963a-c5ef798a569b-kube-api-access-xtqdk\") pod \"packageserver-d55dfcdfc-rsxc6\" (UID: \"6a6c3149-2d76-48de-963a-c5ef798a569b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rsxc6" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.602810 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.612560 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jsrp4\" (UniqueName: \"kubernetes.io/projected/3164dad9-48a2-463c-a2c7-c668eec1ac77-kube-api-access-jsrp4\") pod \"dns-operator-744455d44c-wxsv2\" (UID: \"3164dad9-48a2-463c-a2c7-c668eec1ac77\") " pod="openshift-dns-operator/dns-operator-744455d44c-wxsv2" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.619236 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 00:08:42 crc kubenswrapper[4967]: E1008 00:08:42.619464 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 00:08:43.119441509 +0000 UTC m=+121.769976168 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.619535 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/bb6ecea9-8eb0-4ff6-a898-5cb24d813314-certs\") pod \"machine-config-server-h786z\" (UID: \"bb6ecea9-8eb0-4ff6-a898-5cb24d813314\") " pod="openshift-machine-config-operator/machine-config-server-h786z" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.619583 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gnznl\" (UniqueName: \"kubernetes.io/projected/25199d70-97f2-4513-a2e4-301c9ce3f85a-kube-api-access-gnznl\") pod \"dns-default-fjz5z\" (UID: \"25199d70-97f2-4513-a2e4-301c9ce3f85a\") " pod="openshift-dns/dns-default-fjz5z" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.619631 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/25199d70-97f2-4513-a2e4-301c9ce3f85a-config-volume\") pod \"dns-default-fjz5z\" (UID: \"25199d70-97f2-4513-a2e4-301c9ce3f85a\") " pod="openshift-dns/dns-default-fjz5z" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.619734 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8d8x5\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.619799 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/25199d70-97f2-4513-a2e4-301c9ce3f85a-metrics-tls\") pod \"dns-default-fjz5z\" (UID: \"25199d70-97f2-4513-a2e4-301c9ce3f85a\") " pod="openshift-dns/dns-default-fjz5z" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.619909 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/bb6ecea9-8eb0-4ff6-a898-5cb24d813314-node-bootstrap-token\") pod \"machine-config-server-h786z\" (UID: \"bb6ecea9-8eb0-4ff6-a898-5cb24d813314\") " pod="openshift-machine-config-operator/machine-config-server-h786z" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.619963 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jx6nz\" (UniqueName: \"kubernetes.io/projected/bb6ecea9-8eb0-4ff6-a898-5cb24d813314-kube-api-access-jx6nz\") pod \"machine-config-server-h786z\" (UID: \"bb6ecea9-8eb0-4ff6-a898-5cb24d813314\") " pod="openshift-machine-config-operator/machine-config-server-h786z" Oct 08 00:08:42 crc kubenswrapper[4967]: E1008 00:08:42.620227 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 00:08:43.120210839 +0000 UTC m=+121.770745488 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8d8x5" (UID: "1f06a5d1-8ff8-4d73-8939-0fac4252ea9c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.626015 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.635355 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9mgd5\" (UniqueName: \"kubernetes.io/projected/af03bd84-939d-4a5e-9336-7f53108cdc33-kube-api-access-9mgd5\") pod \"cluster-samples-operator-665b6dd947-fg895\" (UID: \"af03bd84-939d-4a5e-9336-7f53108cdc33\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fg895" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.650702 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.655613 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2mcc\" (UniqueName: \"kubernetes.io/projected/5f033f1c-0cb3-4dcf-ba2b-79593657b841-kube-api-access-b2mcc\") pod \"kube-storage-version-migrator-operator-b67b599dd-c4cg2\" (UID: \"5f033f1c-0cb3-4dcf-ba2b-79593657b841\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c4cg2" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.665081 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-ljjfc"] Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.708170 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1f06a5d1-8ff8-4d73-8939-0fac4252ea9c-bound-sa-token\") pod \"image-registry-697d97f7c8-8d8x5\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.720997 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 00:08:42 crc kubenswrapper[4967]: E1008 00:08:42.721702 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 00:08:43.221687746 +0000 UTC m=+121.872222385 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.723564 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-ck4pd"] Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.723912 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vpcpf\" (UniqueName: \"kubernetes.io/projected/9fa40840-4cd4-41ea-9c97-8a76e3001eee-kube-api-access-vpcpf\") pod \"olm-operator-6b444d44fb-4kzvf\" (UID: \"9fa40840-4cd4-41ea-9c97-8a76e3001eee\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4kzvf" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.746475 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-b5mzq"] Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.763411 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hkk8l\" (UniqueName: \"kubernetes.io/projected/cd33606b-2d0f-4f4b-bc86-b85ad4f56f83-kube-api-access-hkk8l\") pod \"control-plane-machine-set-operator-78cbb6b69f-lpxcq\" (UID: \"cd33606b-2d0f-4f4b-bc86-b85ad4f56f83\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-lpxcq" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.782636 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6lf6\" (UniqueName: \"kubernetes.io/projected/1f06a5d1-8ff8-4d73-8939-0fac4252ea9c-kube-api-access-b6lf6\") pod \"image-registry-697d97f7c8-8d8x5\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.823339 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8d8x5\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:08:42 crc kubenswrapper[4967]: E1008 00:08:42.824121 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 00:08:43.324038554 +0000 UTC m=+121.974573203 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8d8x5" (UID: "1f06a5d1-8ff8-4d73-8939-0fac4252ea9c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.831539 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-48bwp\" (UniqueName: \"kubernetes.io/projected/bb260fdf-16d6-4771-b7ae-38a669ff2393-kube-api-access-48bwp\") pod \"collect-profiles-29331360-t4jlf\" (UID: \"bb260fdf-16d6-4771-b7ae-38a669ff2393\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331360-t4jlf" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.840776 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-64h5g\" (UniqueName: \"kubernetes.io/projected/b2693055-f9b3-4539-ac6d-2151aafe6c2e-kube-api-access-64h5g\") pod \"machine-config-controller-84d6567774-25pkg\" (UID: \"b2693055-f9b3-4539-ac6d-2151aafe6c2e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-25pkg" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.841003 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-p9jtz"] Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.874959 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-b5mzq" event={"ID":"6801f4d5-5cce-4276-833a-20c81735e3be","Type":"ContainerStarted","Data":"7b70630c7343ed1945c80c0f684567d1107c62c3b3da1855726e8370c84296df"} Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.892340 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-ljjfc" event={"ID":"1a14c5f6-9dfd-47fd-acf7-65ec6bbd5149","Type":"ContainerStarted","Data":"c31a43b9c4c2d70d4c1ce861d53645a77bd387b5d78d354290465c250a95b0a1"} Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.896637 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Oct 08 00:08:42 crc kubenswrapper[4967]: W1008 00:08:42.910568 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod86941755_9852_4757_82b6_c4aa2c3ffa43.slice/crio-a8d864775e93384ccf97ccc488ebf9446959f2e59cd053e4e2b875583b88b4dd WatchSource:0}: Error finding container a8d864775e93384ccf97ccc488ebf9446959f2e59cd053e4e2b875583b88b4dd: Status 404 returned error can't find the container with id a8d864775e93384ccf97ccc488ebf9446959f2e59cd053e4e2b875583b88b4dd Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.911586 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f79e3965-8fdc-48ba-be6c-2866348f2458-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-npdfq\" (UID: \"f79e3965-8fdc-48ba-be6c-2866348f2458\") " pod="openshift-marketplace/marketplace-operator-79b997595-npdfq" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.914536 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-ck4pd" event={"ID":"71b5b4fb-2e90-4b39-b976-6806ff38ed8e","Type":"ContainerStarted","Data":"31f4dfd1beac0de59a0f52e27bba96833143fbfb01045e619d857b8052f84b38"} Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.915437 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-pruner-29331360-494rn"] Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.916607 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-krqks" event={"ID":"fccf1013-b838-4303-8da2-4861e234540b","Type":"ContainerStarted","Data":"f356173fe877f71f38805ef729ab177333a77d13af722e4471c14f773fe06718"} Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.916784 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-krqks" event={"ID":"fccf1013-b838-4303-8da2-4861e234540b","Type":"ContainerStarted","Data":"db5c57086329a1b8df15856f2e0ae83c164861f24d72d8853031ad4e24fbba4d"} Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.923225 4967 generic.go:334] "Generic (PLEG): container finished" podID="87f9f2fd-139e-4ac0-a51c-8473ae6f5986" containerID="b96b17a8d68da27a3a2fc7738f98a9c0d3aa67466aa9fff78e7e8a90c65aa0ea" exitCode=0 Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.923294 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-jjhb5" event={"ID":"87f9f2fd-139e-4ac0-a51c-8473ae6f5986","Type":"ContainerDied","Data":"b96b17a8d68da27a3a2fc7738f98a9c0d3aa67466aa9fff78e7e8a90c65aa0ea"} Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.923323 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-jjhb5" event={"ID":"87f9f2fd-139e-4ac0-a51c-8473ae6f5986","Type":"ContainerStarted","Data":"e0fa31c0201f6cf10f7c2567052be99b049c98ef62926d128b090f96d7158fda"} Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.924220 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.927025 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e873b76e-e526-4a38-a2d2-9b23ac147852-config\") pod \"etcd-operator-b45778765-fqmxp\" (UID: \"e873b76e-e526-4a38-a2d2-9b23ac147852\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fqmxp" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.927303 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-tdqfv" event={"ID":"fa677745-a987-479c-b60a-2220f1bd58d1","Type":"ContainerStarted","Data":"7a1796cbfc59729027adce228787547a56eecf206f086705bfb205cc0c6e2225"} Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.927341 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-tdqfv" event={"ID":"fa677745-a987-479c-b60a-2220f1bd58d1","Type":"ContainerStarted","Data":"551e116fdea35ec6dd300bbbf8a10d7a7bcc715f4c260c8bac3deb26e236c504"} Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.928115 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-tdqfv" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.929753 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 00:08:42 crc kubenswrapper[4967]: E1008 00:08:42.931836 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 00:08:43.431813697 +0000 UTC m=+122.082348346 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.938584 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-tmj7c" event={"ID":"d7ec4b5e-d6b4-4681-a231-c7811661c6cb","Type":"ContainerStarted","Data":"30495094bf636d17573cf31907c204329ec27e29d279878d6eeecc6249b8a486"} Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.938677 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-tmj7c" event={"ID":"d7ec4b5e-d6b4-4681-a231-c7811661c6cb","Type":"ContainerStarted","Data":"5f315a5ae87bfac31e4598e500fd632a163a51d5846d1541929ce7ca88ed7a1b"} Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.939927 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-tmj7c" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.944976 4967 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-tmj7c container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.12:6443/healthz\": dial tcp 10.217.0.12:6443: connect: connection refused" start-of-body= Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.945079 4967 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-tmj7c" podUID="d7ec4b5e-d6b4-4681-a231-c7811661c6cb" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.12:6443/healthz\": dial tcp 10.217.0.12:6443: connect: connection refused" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.945216 4967 patch_prober.go:28] interesting pod/downloads-7954f5f757-tdqfv container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.18:8080/\": dial tcp 10.217.0.18:8080: connect: connection refused" start-of-body= Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.945347 4967 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-tdqfv" podUID="fa677745-a987-479c-b60a-2220f1bd58d1" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.18:8080/\": dial tcp 10.217.0.18:8080: connect: connection refused" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.949491 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-tlvk7" event={"ID":"d74a2ad0-73e6-4f0d-a804-b7ede2cf8835","Type":"ContainerStarted","Data":"406adb899327c53ee5365bec990c16df99b697c01326b1342ae1cf06de3b4780"} Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.949540 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-tlvk7" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.949565 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-tlvk7" event={"ID":"d74a2ad0-73e6-4f0d-a804-b7ede2cf8835","Type":"ContainerStarted","Data":"011205476e92f6866751843db10415bb8fb22a084933e0adfc85f99053dd1198"} Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.950905 4967 patch_prober.go:28] interesting pod/console-operator-58897d9998-tlvk7 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.11:8443/readyz\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.951003 4967 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-tlvk7" podUID="d74a2ad0-73e6-4f0d-a804-b7ede2cf8835" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.11:8443/readyz\": dial tcp 10.217.0.11:8443: connect: connection refused" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.952157 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.958349 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8d8x5\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:08:42 crc kubenswrapper[4967]: E1008 00:08:42.962289 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 00:08:43.462265383 +0000 UTC m=+122.112800202 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8d8x5" (UID: "1f06a5d1-8ff8-4d73-8939-0fac4252ea9c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.965384 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v4kxw" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.965443 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-vkf46" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.967264 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-rjwhx"] Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.968986 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4481eaf4-ecf4-46d3-be86-f028e4e33970-trusted-ca\") pod \"ingress-operator-5b745b69d9-wp2tt\" (UID: \"4481eaf4-ecf4-46d3-be86-f028e4e33970\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wp2tt" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.969975 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Oct 08 00:08:42 crc kubenswrapper[4967]: I1008 00:08:42.979421 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/e0d38c54-b3a4-46b1-af38-0f8c7d3dbf5f-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-b5glr\" (UID: \"e0d38c54-b3a4-46b1-af38-0f8c7d3dbf5f\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b5glr" Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.009501 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.013900 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.016085 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/c7b7dad0-eb2c-4c7b-ac09-48d58ed7b936-images\") pod \"machine-config-operator-74547568cd-2ghtr\" (UID: \"c7b7dad0-eb2c-4c7b-ac09-48d58ed7b936\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2ghtr" Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.026534 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f79e3965-8fdc-48ba-be6c-2866348f2458-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-npdfq\" (UID: \"f79e3965-8fdc-48ba-be6c-2866348f2458\") " pod="openshift-marketplace/marketplace-operator-79b997595-npdfq" Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.027702 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.038662 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bb260fdf-16d6-4771-b7ae-38a669ff2393-config-volume\") pod \"collect-profiles-29331360-t4jlf\" (UID: \"bb260fdf-16d6-4771-b7ae-38a669ff2393\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331360-t4jlf" Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.041853 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.047012 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ee7dadb3-3ab7-4790-9b67-7f7a5bf7604f-service-ca-bundle\") pod \"router-default-5444994796-p9kxf\" (UID: \"ee7dadb3-3ab7-4790-9b67-7f7a5bf7604f\") " pod="openshift-ingress/router-default-5444994796-p9kxf" Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.060981 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 00:08:43 crc kubenswrapper[4967]: E1008 00:08:43.062502 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 00:08:43.562487088 +0000 UTC m=+122.213021727 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.093754 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.099076 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7e86e7c5-1431-4606-bf22-7bc05205e8e8-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-pjlzq\" (UID: \"7e86e7c5-1431-4606-bf22-7bc05205e8e8\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-pjlzq" Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.100158 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1f06a5d1-8ff8-4d73-8939-0fac4252ea9c-trusted-ca\") pod \"image-registry-697d97f7c8-8d8x5\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.114761 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.115352 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/e873b76e-e526-4a38-a2d2-9b23ac147852-etcd-service-ca\") pod \"etcd-operator-b45778765-fqmxp\" (UID: \"e873b76e-e526-4a38-a2d2-9b23ac147852\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fqmxp" Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.143212 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lbttt\" (UniqueName: \"kubernetes.io/projected/8e5ce93e-573e-456c-a08c-4737fb3c7a66-kube-api-access-lbttt\") pod \"console-f9d7485db-vxdxg\" (UID: \"8e5ce93e-573e-456c-a08c-4737fb3c7a66\") " pod="openshift-console/console-f9d7485db-vxdxg" Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.146989 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.156400 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/8e5ce93e-573e-456c-a08c-4737fb3c7a66-console-config\") pod \"console-f9d7485db-vxdxg\" (UID: \"8e5ce93e-573e-456c-a08c-4737fb3c7a66\") " pod="openshift-console/console-f9d7485db-vxdxg" Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.165940 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8d8x5\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:08:43 crc kubenswrapper[4967]: E1008 00:08:43.166331 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 00:08:43.666318623 +0000 UTC m=+122.316853262 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8d8x5" (UID: "1f06a5d1-8ff8-4d73-8939-0fac4252ea9c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.166990 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.177765 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/ee7dadb3-3ab7-4790-9b67-7f7a5bf7604f-default-certificate\") pod \"router-default-5444994796-p9kxf\" (UID: \"ee7dadb3-3ab7-4790-9b67-7f7a5bf7604f\") " pod="openshift-ingress/router-default-5444994796-p9kxf" Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.199412 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hbmqm\" (UniqueName: \"kubernetes.io/projected/e0d38c54-b3a4-46b1-af38-0f8c7d3dbf5f-kube-api-access-hbmqm\") pod \"package-server-manager-789f6589d5-b5glr\" (UID: \"e0d38c54-b3a4-46b1-af38-0f8c7d3dbf5f\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b5glr" Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.202778 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.208276 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/c7b7dad0-eb2c-4c7b-ac09-48d58ed7b936-proxy-tls\") pod \"machine-config-operator-74547568cd-2ghtr\" (UID: \"c7b7dad0-eb2c-4c7b-ac09-48d58ed7b936\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2ghtr" Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.247271 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.255554 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8e5ce93e-573e-456c-a08c-4737fb3c7a66-trusted-ca-bundle\") pod \"console-f9d7485db-vxdxg\" (UID: \"8e5ce93e-573e-456c-a08c-4737fb3c7a66\") " pod="openshift-console/console-f9d7485db-vxdxg" Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.266868 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 00:08:43 crc kubenswrapper[4967]: E1008 00:08:43.266989 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 00:08:43.766963729 +0000 UTC m=+122.417498378 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.267406 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8d8x5\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:08:43 crc kubenswrapper[4967]: E1008 00:08:43.267741 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 00:08:43.767733629 +0000 UTC m=+122.418268278 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8d8x5" (UID: "1f06a5d1-8ff8-4d73-8939-0fac4252ea9c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.282352 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vpr4c\" (UniqueName: \"kubernetes.io/projected/c7b7dad0-eb2c-4c7b-ac09-48d58ed7b936-kube-api-access-vpr4c\") pod \"machine-config-operator-74547568cd-2ghtr\" (UID: \"c7b7dad0-eb2c-4c7b-ac09-48d58ed7b936\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2ghtr" Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.283417 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.286776 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/e873b76e-e526-4a38-a2d2-9b23ac147852-etcd-ca\") pod \"etcd-operator-b45778765-fqmxp\" (UID: \"e873b76e-e526-4a38-a2d2-9b23ac147852\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fqmxp" Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.337648 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4481eaf4-ecf4-46d3-be86-f028e4e33970-bound-sa-token\") pod \"ingress-operator-5b745b69d9-wp2tt\" (UID: \"4481eaf4-ecf4-46d3-be86-f028e4e33970\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wp2tt" Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.344124 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.361512 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b2693055-f9b3-4539-ac6d-2151aafe6c2e-proxy-tls\") pod \"machine-config-controller-84d6567774-25pkg\" (UID: \"b2693055-f9b3-4539-ac6d-2151aafe6c2e\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-25pkg" Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.362757 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.366575 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/2c9d9d1d-3e68-47b5-b649-f193653c1922-signing-cabundle\") pod \"service-ca-9c57cc56f-ghwrl\" (UID: \"2c9d9d1d-3e68-47b5-b649-f193653c1922\") " pod="openshift-service-ca/service-ca-9c57cc56f-ghwrl" Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.368062 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 00:08:43 crc kubenswrapper[4967]: E1008 00:08:43.368652 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 00:08:43.868637702 +0000 UTC m=+122.519172351 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.382482 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.396953 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e873b76e-e526-4a38-a2d2-9b23ac147852-etcd-client\") pod \"etcd-operator-b45778765-fqmxp\" (UID: \"e873b76e-e526-4a38-a2d2-9b23ac147852\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fqmxp" Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.416543 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7e86e7c5-1431-4606-bf22-7bc05205e8e8-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-pjlzq\" (UID: \"7e86e7c5-1431-4606-bf22-7bc05205e8e8\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-pjlzq" Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.423085 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.426816 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e47422e6-8d98-4115-b59d-615218bfd4ae-config\") pod \"service-ca-operator-777779d784-kqmk8\" (UID: \"e47422e6-8d98-4115-b59d-615218bfd4ae\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-kqmk8" Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.456695 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9htfq\" (UniqueName: \"kubernetes.io/projected/7e86e7c5-1431-4606-bf22-7bc05205e8e8-kube-api-access-9htfq\") pod \"cluster-image-registry-operator-dc59b4c8b-pjlzq\" (UID: \"7e86e7c5-1431-4606-bf22-7bc05205e8e8\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-pjlzq" Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.462060 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.468834 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ee7dadb3-3ab7-4790-9b67-7f7a5bf7604f-metrics-certs\") pod \"router-default-5444994796-p9kxf\" (UID: \"ee7dadb3-3ab7-4790-9b67-7f7a5bf7604f\") " pod="openshift-ingress/router-default-5444994796-p9kxf" Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.469617 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8d8x5\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:08:43 crc kubenswrapper[4967]: E1008 00:08:43.470144 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 00:08:43.970125929 +0000 UTC m=+122.620660578 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8d8x5" (UID: "1f06a5d1-8ff8-4d73-8939-0fac4252ea9c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.486745 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.500226 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/4481eaf4-ecf4-46d3-be86-f028e4e33970-metrics-tls\") pod \"ingress-operator-5b745b69d9-wp2tt\" (UID: \"4481eaf4-ecf4-46d3-be86-f028e4e33970\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wp2tt" Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.502736 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.509428 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/2c9d9d1d-3e68-47b5-b649-f193653c1922-signing-key\") pod \"service-ca-9c57cc56f-ghwrl\" (UID: \"2c9d9d1d-3e68-47b5-b649-f193653c1922\") " pod="openshift-service-ca/service-ca-9c57cc56f-ghwrl" Oct 08 00:08:43 crc kubenswrapper[4967]: E1008 00:08:43.515378 4967 projected.go:263] Couldn't get secret openshift-image-registry/image-registry-tls: failed to sync secret cache: timed out waiting for the condition Oct 08 00:08:43 crc kubenswrapper[4967]: E1008 00:08:43.515429 4967 projected.go:194] Error preparing data for projected volume registry-tls for pod openshift-image-registry/image-registry-697d97f7c8-8d8x5: failed to sync secret cache: timed out waiting for the condition Oct 08 00:08:43 crc kubenswrapper[4967]: E1008 00:08:43.515541 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/1f06a5d1-8ff8-4d73-8939-0fac4252ea9c-registry-tls podName:1f06a5d1-8ff8-4d73-8939-0fac4252ea9c nodeName:}" failed. No retries permitted until 2025-10-08 00:08:44.015487724 +0000 UTC m=+122.666022373 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "registry-tls" (UniqueName: "kubernetes.io/projected/1f06a5d1-8ff8-4d73-8939-0fac4252ea9c-registry-tls") pod "image-registry-697d97f7c8-8d8x5" (UID: "1f06a5d1-8ff8-4d73-8939-0fac4252ea9c") : failed to sync secret cache: timed out waiting for the condition Oct 08 00:08:43 crc kubenswrapper[4967]: E1008 00:08:43.515558 4967 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/olm-operator-serving-cert: failed to sync secret cache: timed out waiting for the condition Oct 08 00:08:43 crc kubenswrapper[4967]: E1008 00:08:43.515747 4967 secret.go:188] Couldn't get secret openshift-ingress-canary/canary-serving-cert: failed to sync secret cache: timed out waiting for the condition Oct 08 00:08:43 crc kubenswrapper[4967]: E1008 00:08:43.515787 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9fa40840-4cd4-41ea-9c97-8a76e3001eee-srv-cert podName:9fa40840-4cd4-41ea-9c97-8a76e3001eee nodeName:}" failed. No retries permitted until 2025-10-08 00:08:44.015656288 +0000 UTC m=+122.666190937 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "srv-cert" (UniqueName: "kubernetes.io/secret/9fa40840-4cd4-41ea-9c97-8a76e3001eee-srv-cert") pod "olm-operator-6b444d44fb-4kzvf" (UID: "9fa40840-4cd4-41ea-9c97-8a76e3001eee") : failed to sync secret cache: timed out waiting for the condition Oct 08 00:08:43 crc kubenswrapper[4967]: E1008 00:08:43.515810 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/64202fa4-cb80-4028-aa03-d53231f2d8f4-cert podName:64202fa4-cb80-4028-aa03-d53231f2d8f4 nodeName:}" failed. No retries permitted until 2025-10-08 00:08:44.015801682 +0000 UTC m=+122.666336331 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/64202fa4-cb80-4028-aa03-d53231f2d8f4-cert") pod "ingress-canary-hpx62" (UID: "64202fa4-cb80-4028-aa03-d53231f2d8f4") : failed to sync secret cache: timed out waiting for the condition Oct 08 00:08:43 crc kubenswrapper[4967]: E1008 00:08:43.515844 4967 secret.go:188] Couldn't get secret openshift-etcd-operator/etcd-operator-serving-cert: failed to sync secret cache: timed out waiting for the condition Oct 08 00:08:43 crc kubenswrapper[4967]: E1008 00:08:43.515882 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e873b76e-e526-4a38-a2d2-9b23ac147852-serving-cert podName:e873b76e-e526-4a38-a2d2-9b23ac147852 nodeName:}" failed. No retries permitted until 2025-10-08 00:08:44.015868703 +0000 UTC m=+122.666403352 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/e873b76e-e526-4a38-a2d2-9b23ac147852-serving-cert") pod "etcd-operator-b45778765-fqmxp" (UID: "e873b76e-e526-4a38-a2d2-9b23ac147852") : failed to sync secret cache: timed out waiting for the condition Oct 08 00:08:43 crc kubenswrapper[4967]: E1008 00:08:43.518210 4967 secret.go:188] Couldn't get secret openshift-ingress/router-stats-default: failed to sync secret cache: timed out waiting for the condition Oct 08 00:08:43 crc kubenswrapper[4967]: E1008 00:08:43.518261 4967 secret.go:188] Couldn't get secret openshift-console/console-oauth-config: failed to sync secret cache: timed out waiting for the condition Oct 08 00:08:43 crc kubenswrapper[4967]: E1008 00:08:43.518312 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8e5ce93e-573e-456c-a08c-4737fb3c7a66-console-oauth-config podName:8e5ce93e-573e-456c-a08c-4737fb3c7a66 nodeName:}" failed. No retries permitted until 2025-10-08 00:08:44.018297714 +0000 UTC m=+122.668832603 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "console-oauth-config" (UniqueName: "kubernetes.io/secret/8e5ce93e-573e-456c-a08c-4737fb3c7a66-console-oauth-config") pod "console-f9d7485db-vxdxg" (UID: "8e5ce93e-573e-456c-a08c-4737fb3c7a66") : failed to sync secret cache: timed out waiting for the condition Oct 08 00:08:43 crc kubenswrapper[4967]: E1008 00:08:43.518378 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ee7dadb3-3ab7-4790-9b67-7f7a5bf7604f-stats-auth podName:ee7dadb3-3ab7-4790-9b67-7f7a5bf7604f nodeName:}" failed. No retries permitted until 2025-10-08 00:08:44.018368975 +0000 UTC m=+122.668903864 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "stats-auth" (UniqueName: "kubernetes.io/secret/ee7dadb3-3ab7-4790-9b67-7f7a5bf7604f-stats-auth") pod "router-default-5444994796-p9kxf" (UID: "ee7dadb3-3ab7-4790-9b67-7f7a5bf7604f") : failed to sync secret cache: timed out waiting for the condition Oct 08 00:08:43 crc kubenswrapper[4967]: E1008 00:08:43.519492 4967 secret.go:188] Couldn't get secret openshift-image-registry/image-registry-operator-tls: failed to sync secret cache: timed out waiting for the condition Oct 08 00:08:43 crc kubenswrapper[4967]: E1008 00:08:43.519517 4967 configmap.go:193] Couldn't get configMap openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-config: failed to sync configmap cache: timed out waiting for the condition Oct 08 00:08:43 crc kubenswrapper[4967]: E1008 00:08:43.519566 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/c2302f69-8556-43c8-9451-145211206874-config podName:c2302f69-8556-43c8-9451-145211206874 nodeName:}" failed. No retries permitted until 2025-10-08 00:08:44.019553425 +0000 UTC m=+122.670088294 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/c2302f69-8556-43c8-9451-145211206874-config") pod "openshift-kube-scheduler-operator-5fdd9b5758-n69qz" (UID: "c2302f69-8556-43c8-9451-145211206874") : failed to sync configmap cache: timed out waiting for the condition Oct 08 00:08:43 crc kubenswrapper[4967]: E1008 00:08:43.519567 4967 secret.go:188] Couldn't get secret openshift-image-registry/installation-pull-secrets: failed to sync secret cache: timed out waiting for the condition Oct 08 00:08:43 crc kubenswrapper[4967]: E1008 00:08:43.519588 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7e86e7c5-1431-4606-bf22-7bc05205e8e8-image-registry-operator-tls podName:7e86e7c5-1431-4606-bf22-7bc05205e8e8 nodeName:}" failed. No retries permitted until 2025-10-08 00:08:44.019575786 +0000 UTC m=+122.670110685 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "image-registry-operator-tls" (UniqueName: "kubernetes.io/secret/7e86e7c5-1431-4606-bf22-7bc05205e8e8-image-registry-operator-tls") pod "cluster-image-registry-operator-dc59b4c8b-pjlzq" (UID: "7e86e7c5-1431-4606-bf22-7bc05205e8e8") : failed to sync secret cache: timed out waiting for the condition Oct 08 00:08:43 crc kubenswrapper[4967]: E1008 00:08:43.519619 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1f06a5d1-8ff8-4d73-8939-0fac4252ea9c-installation-pull-secrets podName:1f06a5d1-8ff8-4d73-8939-0fac4252ea9c nodeName:}" failed. No retries permitted until 2025-10-08 00:08:44.019596876 +0000 UTC m=+122.670131525 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "installation-pull-secrets" (UniqueName: "kubernetes.io/secret/1f06a5d1-8ff8-4d73-8939-0fac4252ea9c-installation-pull-secrets") pod "image-registry-697d97f7c8-8d8x5" (UID: "1f06a5d1-8ff8-4d73-8939-0fac4252ea9c") : failed to sync secret cache: timed out waiting for the condition Oct 08 00:08:43 crc kubenswrapper[4967]: E1008 00:08:43.519621 4967 configmap.go:193] Couldn't get configMap openshift-console/oauth-serving-cert: failed to sync configmap cache: timed out waiting for the condition Oct 08 00:08:43 crc kubenswrapper[4967]: E1008 00:08:43.519659 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/8e5ce93e-573e-456c-a08c-4737fb3c7a66-oauth-serving-cert podName:8e5ce93e-573e-456c-a08c-4737fb3c7a66 nodeName:}" failed. No retries permitted until 2025-10-08 00:08:44.019652928 +0000 UTC m=+122.670187577 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "oauth-serving-cert" (UniqueName: "kubernetes.io/configmap/8e5ce93e-573e-456c-a08c-4737fb3c7a66-oauth-serving-cert") pod "console-f9d7485db-vxdxg" (UID: "8e5ce93e-573e-456c-a08c-4737fb3c7a66") : failed to sync configmap cache: timed out waiting for the condition Oct 08 00:08:43 crc kubenswrapper[4967]: E1008 00:08:43.519681 4967 configmap.go:193] Couldn't get configMap openshift-console/service-ca: failed to sync configmap cache: timed out waiting for the condition Oct 08 00:08:43 crc kubenswrapper[4967]: E1008 00:08:43.519698 4967 secret.go:188] Couldn't get secret openshift-kube-scheduler-operator/kube-scheduler-operator-serving-cert: failed to sync secret cache: timed out waiting for the condition Oct 08 00:08:43 crc kubenswrapper[4967]: E1008 00:08:43.519715 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/8e5ce93e-573e-456c-a08c-4737fb3c7a66-service-ca podName:8e5ce93e-573e-456c-a08c-4737fb3c7a66 nodeName:}" failed. No retries permitted until 2025-10-08 00:08:44.019706249 +0000 UTC m=+122.670241148 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "service-ca" (UniqueName: "kubernetes.io/configmap/8e5ce93e-573e-456c-a08c-4737fb3c7a66-service-ca") pod "console-f9d7485db-vxdxg" (UID: "8e5ce93e-573e-456c-a08c-4737fb3c7a66") : failed to sync configmap cache: timed out waiting for the condition Oct 08 00:08:43 crc kubenswrapper[4967]: E1008 00:08:43.519745 4967 secret.go:188] Couldn't get secret openshift-console/console-serving-cert: failed to sync secret cache: timed out waiting for the condition Oct 08 00:08:43 crc kubenswrapper[4967]: E1008 00:08:43.519755 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c2302f69-8556-43c8-9451-145211206874-serving-cert podName:c2302f69-8556-43c8-9451-145211206874 nodeName:}" failed. No retries permitted until 2025-10-08 00:08:44.0197488 +0000 UTC m=+122.670283449 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/c2302f69-8556-43c8-9451-145211206874-serving-cert") pod "openshift-kube-scheduler-operator-5fdd9b5758-n69qz" (UID: "c2302f69-8556-43c8-9451-145211206874") : failed to sync secret cache: timed out waiting for the condition Oct 08 00:08:43 crc kubenswrapper[4967]: E1008 00:08:43.519830 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8e5ce93e-573e-456c-a08c-4737fb3c7a66-console-serving-cert podName:8e5ce93e-573e-456c-a08c-4737fb3c7a66 nodeName:}" failed. No retries permitted until 2025-10-08 00:08:44.019818962 +0000 UTC m=+122.670353821 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "console-serving-cert" (UniqueName: "kubernetes.io/secret/8e5ce93e-573e-456c-a08c-4737fb3c7a66-console-serving-cert") pod "console-f9d7485db-vxdxg" (UID: "8e5ce93e-573e-456c-a08c-4737fb3c7a66") : failed to sync secret cache: timed out waiting for the condition Oct 08 00:08:43 crc kubenswrapper[4967]: E1008 00:08:43.519898 4967 secret.go:188] Couldn't get secret openshift-service-ca-operator/serving-cert: failed to sync secret cache: timed out waiting for the condition Oct 08 00:08:43 crc kubenswrapper[4967]: E1008 00:08:43.519941 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e47422e6-8d98-4115-b59d-615218bfd4ae-serving-cert podName:e47422e6-8d98-4115-b59d-615218bfd4ae nodeName:}" failed. No retries permitted until 2025-10-08 00:08:44.019928034 +0000 UTC m=+122.670462923 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/e47422e6-8d98-4115-b59d-615218bfd4ae-serving-cert") pod "service-ca-operator-777779d784-kqmk8" (UID: "e47422e6-8d98-4115-b59d-615218bfd4ae") : failed to sync secret cache: timed out waiting for the condition Oct 08 00:08:43 crc kubenswrapper[4967]: E1008 00:08:43.520002 4967 secret.go:188] Couldn't get secret openshift-machine-api/control-plane-machine-set-operator-tls: failed to sync secret cache: timed out waiting for the condition Oct 08 00:08:43 crc kubenswrapper[4967]: E1008 00:08:43.520033 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cd33606b-2d0f-4f4b-bc86-b85ad4f56f83-control-plane-machine-set-operator-tls podName:cd33606b-2d0f-4f4b-bc86-b85ad4f56f83 nodeName:}" failed. No retries permitted until 2025-10-08 00:08:44.020024207 +0000 UTC m=+122.670559096 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "control-plane-machine-set-operator-tls" (UniqueName: "kubernetes.io/secret/cd33606b-2d0f-4f4b-bc86-b85ad4f56f83-control-plane-machine-set-operator-tls") pod "control-plane-machine-set-operator-78cbb6b69f-lpxcq" (UID: "cd33606b-2d0f-4f4b-bc86-b85ad4f56f83") : failed to sync secret cache: timed out waiting for the condition Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.521192 4967 request.go:700] Waited for 1.005030918s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-operator-lifecycle-manager/secrets?fieldSelector=metadata.name%3Dolm-operator-serving-cert&resourceVersion=27139 Oct 08 00:08:43 crc kubenswrapper[4967]: E1008 00:08:43.522094 4967 secret.go:188] Couldn't get secret openshift-multus/multus-admission-controller-secret: failed to sync secret cache: timed out waiting for the condition Oct 08 00:08:43 crc kubenswrapper[4967]: E1008 00:08:43.522146 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/553037a9-2f32-4eee-8244-caa084e58566-webhook-certs podName:553037a9-2f32-4eee-8244-caa084e58566 nodeName:}" failed. No retries permitted until 2025-10-08 00:08:44.022132438 +0000 UTC m=+122.672667187 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/553037a9-2f32-4eee-8244-caa084e58566-webhook-certs") pod "multus-admission-controller-857f4d67dd-86lvz" (UID: "553037a9-2f32-4eee-8244-caa084e58566") : failed to sync secret cache: timed out waiting for the condition Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.522312 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.542225 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.563232 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.571300 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 00:08:43 crc kubenswrapper[4967]: E1008 00:08:43.571429 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 00:08:44.071405041 +0000 UTC m=+122.721939700 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.571698 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8d8x5\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:08:43 crc kubenswrapper[4967]: E1008 00:08:43.572139 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 00:08:44.072128599 +0000 UTC m=+122.722663258 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8d8x5" (UID: "1f06a5d1-8ff8-4d73-8939-0fac4252ea9c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.582343 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.594154 4967 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ltttx" secret="" err="failed to sync secret cache: timed out waiting for the condition" Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.594215 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ltttx" Oct 08 00:08:43 crc kubenswrapper[4967]: E1008 00:08:43.620415 4967 secret.go:188] Couldn't get secret openshift-machine-config-operator/machine-config-server-tls: failed to sync secret cache: timed out waiting for the condition Oct 08 00:08:43 crc kubenswrapper[4967]: E1008 00:08:43.620468 4967 secret.go:188] Couldn't get secret openshift-machine-config-operator/node-bootstrapper-token: failed to sync secret cache: timed out waiting for the condition Oct 08 00:08:43 crc kubenswrapper[4967]: E1008 00:08:43.620616 4967 configmap.go:193] Couldn't get configMap openshift-dns/dns-default: failed to sync configmap cache: timed out waiting for the condition Oct 08 00:08:43 crc kubenswrapper[4967]: E1008 00:08:43.620524 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bb6ecea9-8eb0-4ff6-a898-5cb24d813314-certs podName:bb6ecea9-8eb0-4ff6-a898-5cb24d813314 nodeName:}" failed. No retries permitted until 2025-10-08 00:08:44.120498989 +0000 UTC m=+122.771033648 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "certs" (UniqueName: "kubernetes.io/secret/bb6ecea9-8eb0-4ff6-a898-5cb24d813314-certs") pod "machine-config-server-h786z" (UID: "bb6ecea9-8eb0-4ff6-a898-5cb24d813314") : failed to sync secret cache: timed out waiting for the condition Oct 08 00:08:43 crc kubenswrapper[4967]: E1008 00:08:43.620646 4967 secret.go:188] Couldn't get secret openshift-dns/dns-default-metrics-tls: failed to sync secret cache: timed out waiting for the condition Oct 08 00:08:43 crc kubenswrapper[4967]: E1008 00:08:43.620716 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bb6ecea9-8eb0-4ff6-a898-5cb24d813314-node-bootstrap-token podName:bb6ecea9-8eb0-4ff6-a898-5cb24d813314 nodeName:}" failed. No retries permitted until 2025-10-08 00:08:44.120662513 +0000 UTC m=+122.771197372 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "node-bootstrap-token" (UniqueName: "kubernetes.io/secret/bb6ecea9-8eb0-4ff6-a898-5cb24d813314-node-bootstrap-token") pod "machine-config-server-h786z" (UID: "bb6ecea9-8eb0-4ff6-a898-5cb24d813314") : failed to sync secret cache: timed out waiting for the condition Oct 08 00:08:43 crc kubenswrapper[4967]: E1008 00:08:43.620740 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/25199d70-97f2-4513-a2e4-301c9ce3f85a-config-volume podName:25199d70-97f2-4513-a2e4-301c9ce3f85a nodeName:}" failed. No retries permitted until 2025-10-08 00:08:44.120731695 +0000 UTC m=+122.771266344 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-volume" (UniqueName: "kubernetes.io/configmap/25199d70-97f2-4513-a2e4-301c9ce3f85a-config-volume") pod "dns-default-fjz5z" (UID: "25199d70-97f2-4513-a2e4-301c9ce3f85a") : failed to sync configmap cache: timed out waiting for the condition Oct 08 00:08:43 crc kubenswrapper[4967]: E1008 00:08:43.620858 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/25199d70-97f2-4513-a2e4-301c9ce3f85a-metrics-tls podName:25199d70-97f2-4513-a2e4-301c9ce3f85a nodeName:}" failed. No retries permitted until 2025-10-08 00:08:44.120761715 +0000 UTC m=+122.771296364 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/25199d70-97f2-4513-a2e4-301c9ce3f85a-metrics-tls") pod "dns-default-fjz5z" (UID: "25199d70-97f2-4513-a2e4-301c9ce3f85a") : failed to sync secret cache: timed out waiting for the condition Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.623037 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.642845 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.656954 4967 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-tzhwk" secret="" err="failed to sync secret cache: timed out waiting for the condition" Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.657045 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-tzhwk" Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.662886 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.673334 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 00:08:43 crc kubenswrapper[4967]: E1008 00:08:43.673942 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 00:08:44.173917014 +0000 UTC m=+122.824451663 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.688384 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.704270 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.722587 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.735889 4967 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c4cg2" secret="" err="failed to sync secret cache: timed out waiting for the condition" Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.735979 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c4cg2" Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.737027 4967 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cpnmt" secret="" err="failed to sync secret cache: timed out waiting for the condition" Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.737075 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cpnmt" Oct 08 00:08:43 crc kubenswrapper[4967]: E1008 00:08:43.740918 4967 projected.go:288] Couldn't get configMap openshift-kube-scheduler-operator/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Oct 08 00:08:43 crc kubenswrapper[4967]: E1008 00:08:43.740960 4967 projected.go:194] Error preparing data for projected volume kube-api-access for pod openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n69qz: failed to sync configmap cache: timed out waiting for the condition Oct 08 00:08:43 crc kubenswrapper[4967]: E1008 00:08:43.741028 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c2302f69-8556-43c8-9451-145211206874-kube-api-access podName:c2302f69-8556-43c8-9451-145211206874 nodeName:}" failed. No retries permitted until 2025-10-08 00:08:44.241000218 +0000 UTC m=+122.891534867 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access" (UniqueName: "kubernetes.io/projected/c2302f69-8556-43c8-9451-145211206874-kube-api-access") pod "openshift-kube-scheduler-operator-5fdd9b5758-n69qz" (UID: "c2302f69-8556-43c8-9451-145211206874") : failed to sync configmap cache: timed out waiting for the condition Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.743130 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.744982 4967 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rsxc6" secret="" err="failed to sync secret cache: timed out waiting for the condition" Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.745084 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rsxc6" Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.767523 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.781264 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8d8x5\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:08:43 crc kubenswrapper[4967]: E1008 00:08:43.781826 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 00:08:44.281813939 +0000 UTC m=+122.932348578 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8d8x5" (UID: "1f06a5d1-8ff8-4d73-8939-0fac4252ea9c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.795080 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.813982 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.835075 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Oct 08 00:08:43 crc kubenswrapper[4967]: E1008 00:08:43.840811 4967 projected.go:288] Couldn't get configMap openshift-etcd-operator/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.863051 4967 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fg895" secret="" err="failed to sync secret cache: timed out waiting for the condition" Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.863102 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fg895" Oct 08 00:08:43 crc kubenswrapper[4967]: E1008 00:08:43.866301 4967 projected.go:288] Couldn't get configMap openshift-ingress-operator/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.872532 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zt77j\" (UniqueName: \"kubernetes.io/projected/553037a9-2f32-4eee-8244-caa084e58566-kube-api-access-zt77j\") pod \"multus-admission-controller-857f4d67dd-86lvz\" (UID: \"553037a9-2f32-4eee-8244-caa084e58566\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-86lvz" Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.888917 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.890448 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 00:08:43 crc kubenswrapper[4967]: E1008 00:08:43.891507 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 00:08:44.391476589 +0000 UTC m=+123.042011238 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:43 crc kubenswrapper[4967]: E1008 00:08:43.902958 4967 projected.go:288] Couldn't get configMap openshift-service-ca/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.902970 4967 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openshift-dns-operator/dns-operator-744455d44c-wxsv2" secret="" err="failed to sync secret cache: timed out waiting for the condition" Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.903078 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-wxsv2" Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.907009 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Oct 08 00:08:43 crc kubenswrapper[4967]: E1008 00:08:43.920210 4967 projected.go:288] Couldn't get configMap hostpath-provisioner/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.921535 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ltttx"] Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.925360 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Oct 08 00:08:43 crc kubenswrapper[4967]: I1008 00:08:43.945235 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.000206 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.002337 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Oct 08 00:08:44 crc kubenswrapper[4967]: E1008 00:08:44.004306 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 00:08:44.504289068 +0000 UTC m=+123.154823717 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8d8x5" (UID: "1f06a5d1-8ff8-4d73-8939-0fac4252ea9c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.016896 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8d8x5\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.017273 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/1f06a5d1-8ff8-4d73-8939-0fac4252ea9c-registry-tls\") pod \"image-registry-697d97f7c8-8d8x5\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.017314 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/9fa40840-4cd4-41ea-9c97-8a76e3001eee-srv-cert\") pod \"olm-operator-6b444d44fb-4kzvf\" (UID: \"9fa40840-4cd4-41ea-9c97-8a76e3001eee\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4kzvf" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.024193 4967 generic.go:334] "Generic (PLEG): container finished" podID="bb075d42-7ef0-4ad4-b817-bdd611f0d7ed" containerID="d86efd8c11ad025872c97e993dccf42d87089467179b32327dd296eac97512cc" exitCode=0 Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.024306 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rjwhx" event={"ID":"bb075d42-7ef0-4ad4-b817-bdd611f0d7ed","Type":"ContainerDied","Data":"d86efd8c11ad025872c97e993dccf42d87089467179b32327dd296eac97512cc"} Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.024335 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rjwhx" event={"ID":"bb075d42-7ef0-4ad4-b817-bdd611f0d7ed","Type":"ContainerStarted","Data":"33e5b328a22b88cf40239f028580f58f2a6a3754a434d2fba5c1202214e1770a"} Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.025259 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/1f06a5d1-8ff8-4d73-8939-0fac4252ea9c-registry-tls\") pod \"image-registry-697d97f7c8-8d8x5\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.035050 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/9fa40840-4cd4-41ea-9c97-8a76e3001eee-srv-cert\") pod \"olm-operator-6b444d44fb-4kzvf\" (UID: \"9fa40840-4cd4-41ea-9c97-8a76e3001eee\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4kzvf" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.036169 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gnznl\" (UniqueName: \"kubernetes.io/projected/25199d70-97f2-4513-a2e4-301c9ce3f85a-kube-api-access-gnznl\") pod \"dns-default-fjz5z\" (UID: \"25199d70-97f2-4513-a2e4-301c9ce3f85a\") " pod="openshift-dns/dns-default-fjz5z" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.037787 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-tzhwk"] Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.045047 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.047225 4967 generic.go:334] "Generic (PLEG): container finished" podID="71b5b4fb-2e90-4b39-b976-6806ff38ed8e" containerID="7adafdcc39ba28dc6adbd43e4d4930c3d59a176997500380d46dc5e45be168e3" exitCode=0 Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.047309 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-ck4pd" event={"ID":"71b5b4fb-2e90-4b39-b976-6806ff38ed8e","Type":"ContainerDied","Data":"7adafdcc39ba28dc6adbd43e4d4930c3d59a176997500380d46dc5e45be168e3"} Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.050430 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jx6nz\" (UniqueName: \"kubernetes.io/projected/bb6ecea9-8eb0-4ff6-a898-5cb24d813314-kube-api-access-jx6nz\") pod \"machine-config-server-h786z\" (UID: \"bb6ecea9-8eb0-4ff6-a898-5cb24d813314\") " pod="openshift-machine-config-operator/machine-config-server-h786z" Oct 08 00:08:44 crc kubenswrapper[4967]: E1008 00:08:44.086909 4967 projected.go:288] Couldn't get configMap openshift-ingress/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.091625 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-ljjfc" event={"ID":"1a14c5f6-9dfd-47fd-acf7-65ec6bbd5149","Type":"ContainerStarted","Data":"6a959cecf09b6f30739f29cf0c5e87d5d6e1a235cdf0d40b47ee5a6727ae94d1"} Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.091705 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-ljjfc" event={"ID":"1a14c5f6-9dfd-47fd-acf7-65ec6bbd5149","Type":"ContainerStarted","Data":"ab4a2e05cbc59b33bf2af6f681fa7b2895fcb54019cf2ef8bd9f38720af4c2de"} Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.097798 4967 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openshift-operator-lifecycle-manager/collect-profiles-29331360-t4jlf" secret="" err="failed to sync secret cache: timed out waiting for the condition" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.097877 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331360-t4jlf" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.106325 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.106605 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.110391 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-jjhb5" event={"ID":"87f9f2fd-139e-4ac0-a51c-8473ae6f5986","Type":"ContainerStarted","Data":"b80d93729dec3463bdd324d3389fdf8493750a3c1d5168b8d5b60001e0b07edd"} Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.110502 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-jjhb5" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.115360 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b5glr" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.118852 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.119102 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c2302f69-8556-43c8-9451-145211206874-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-n69qz\" (UID: \"c2302f69-8556-43c8-9451-145211206874\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n69qz" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.119142 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/7e86e7c5-1431-4606-bf22-7bc05205e8e8-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-pjlzq\" (UID: \"7e86e7c5-1431-4606-bf22-7bc05205e8e8\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-pjlzq" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.119162 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8e5ce93e-573e-456c-a08c-4737fb3c7a66-service-ca\") pod \"console-f9d7485db-vxdxg\" (UID: \"8e5ce93e-573e-456c-a08c-4737fb3c7a66\") " pod="openshift-console/console-f9d7485db-vxdxg" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.119281 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2302f69-8556-43c8-9451-145211206874-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-n69qz\" (UID: \"c2302f69-8556-43c8-9451-145211206874\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n69qz" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.119340 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/1f06a5d1-8ff8-4d73-8939-0fac4252ea9c-installation-pull-secrets\") pod \"image-registry-697d97f7c8-8d8x5\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.119405 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/8e5ce93e-573e-456c-a08c-4737fb3c7a66-console-oauth-config\") pod \"console-f9d7485db-vxdxg\" (UID: \"8e5ce93e-573e-456c-a08c-4737fb3c7a66\") " pod="openshift-console/console-f9d7485db-vxdxg" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.119426 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/8e5ce93e-573e-456c-a08c-4737fb3c7a66-console-serving-cert\") pod \"console-f9d7485db-vxdxg\" (UID: \"8e5ce93e-573e-456c-a08c-4737fb3c7a66\") " pod="openshift-console/console-f9d7485db-vxdxg" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.119441 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/ee7dadb3-3ab7-4790-9b67-7f7a5bf7604f-stats-auth\") pod \"router-default-5444994796-p9kxf\" (UID: \"ee7dadb3-3ab7-4790-9b67-7f7a5bf7604f\") " pod="openshift-ingress/router-default-5444994796-p9kxf" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.119459 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/cd33606b-2d0f-4f4b-bc86-b85ad4f56f83-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-lpxcq\" (UID: \"cd33606b-2d0f-4f4b-bc86-b85ad4f56f83\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-lpxcq" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.119486 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/8e5ce93e-573e-456c-a08c-4737fb3c7a66-oauth-serving-cert\") pod \"console-f9d7485db-vxdxg\" (UID: \"8e5ce93e-573e-456c-a08c-4737fb3c7a66\") " pod="openshift-console/console-f9d7485db-vxdxg" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.119604 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/553037a9-2f32-4eee-8244-caa084e58566-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-86lvz\" (UID: \"553037a9-2f32-4eee-8244-caa084e58566\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-86lvz" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.119627 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e873b76e-e526-4a38-a2d2-9b23ac147852-serving-cert\") pod \"etcd-operator-b45778765-fqmxp\" (UID: \"e873b76e-e526-4a38-a2d2-9b23ac147852\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fqmxp" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.119642 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/64202fa4-cb80-4028-aa03-d53231f2d8f4-cert\") pod \"ingress-canary-hpx62\" (UID: \"64202fa4-cb80-4028-aa03-d53231f2d8f4\") " pod="openshift-ingress-canary/ingress-canary-hpx62" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.119662 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e47422e6-8d98-4115-b59d-615218bfd4ae-serving-cert\") pod \"service-ca-operator-777779d784-kqmk8\" (UID: \"e47422e6-8d98-4115-b59d-615218bfd4ae\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-kqmk8" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.120646 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8e5ce93e-573e-456c-a08c-4737fb3c7a66-service-ca\") pod \"console-f9d7485db-vxdxg\" (UID: \"8e5ce93e-573e-456c-a08c-4737fb3c7a66\") " pod="openshift-console/console-f9d7485db-vxdxg" Oct 08 00:08:44 crc kubenswrapper[4967]: E1008 00:08:44.121951 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 00:08:44.621915556 +0000 UTC m=+123.272450205 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.124197 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c2302f69-8556-43c8-9451-145211206874-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-n69qz\" (UID: \"c2302f69-8556-43c8-9451-145211206874\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n69qz" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.126516 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/8e5ce93e-573e-456c-a08c-4737fb3c7a66-oauth-serving-cert\") pod \"console-f9d7485db-vxdxg\" (UID: \"8e5ce93e-573e-456c-a08c-4737fb3c7a66\") " pod="openshift-console/console-f9d7485db-vxdxg" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.126658 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.138611 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/553037a9-2f32-4eee-8244-caa084e58566-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-86lvz\" (UID: \"553037a9-2f32-4eee-8244-caa084e58566\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-86lvz" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.140200 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/8e5ce93e-573e-456c-a08c-4737fb3c7a66-console-serving-cert\") pod \"console-f9d7485db-vxdxg\" (UID: \"8e5ce93e-573e-456c-a08c-4737fb3c7a66\") " pod="openshift-console/console-f9d7485db-vxdxg" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.142044 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.147443 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e47422e6-8d98-4115-b59d-615218bfd4ae-serving-cert\") pod \"service-ca-operator-777779d784-kqmk8\" (UID: \"e47422e6-8d98-4115-b59d-615218bfd4ae\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-kqmk8" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.150112 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c2302f69-8556-43c8-9451-145211206874-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-n69qz\" (UID: \"c2302f69-8556-43c8-9451-145211206874\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n69qz" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.155446 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/8e5ce93e-573e-456c-a08c-4737fb3c7a66-console-oauth-config\") pod \"console-f9d7485db-vxdxg\" (UID: \"8e5ce93e-573e-456c-a08c-4737fb3c7a66\") " pod="openshift-console/console-f9d7485db-vxdxg" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.157005 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e873b76e-e526-4a38-a2d2-9b23ac147852-serving-cert\") pod \"etcd-operator-b45778765-fqmxp\" (UID: \"e873b76e-e526-4a38-a2d2-9b23ac147852\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fqmxp" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.157310 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/7e86e7c5-1431-4606-bf22-7bc05205e8e8-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-pjlzq\" (UID: \"7e86e7c5-1431-4606-bf22-7bc05205e8e8\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-pjlzq" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.157620 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/1f06a5d1-8ff8-4d73-8939-0fac4252ea9c-installation-pull-secrets\") pod \"image-registry-697d97f7c8-8d8x5\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.158249 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/64202fa4-cb80-4028-aa03-d53231f2d8f4-cert\") pod \"ingress-canary-hpx62\" (UID: \"64202fa4-cb80-4028-aa03-d53231f2d8f4\") " pod="openshift-ingress-canary/ingress-canary-hpx62" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.160006 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/cd33606b-2d0f-4f4b-bc86-b85ad4f56f83-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-lpxcq\" (UID: \"cd33606b-2d0f-4f4b-bc86-b85ad4f56f83\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-lpxcq" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.161556 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/ee7dadb3-3ab7-4790-9b67-7f7a5bf7604f-stats-auth\") pod \"router-default-5444994796-p9kxf\" (UID: \"ee7dadb3-3ab7-4790-9b67-7f7a5bf7604f\") " pod="openshift-ingress/router-default-5444994796-p9kxf" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.168989 4967 patch_prober.go:28] interesting pod/downloads-7954f5f757-tdqfv container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.18:8080/\": dial tcp 10.217.0.18:8080: connect: connection refused" start-of-body= Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.169048 4967 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-tdqfv" podUID="fa677745-a987-479c-b60a-2220f1bd58d1" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.18:8080/\": dial tcp 10.217.0.18:8080: connect: connection refused" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.169206 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.186044 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.203021 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.222985 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.223053 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/bb6ecea9-8eb0-4ff6-a898-5cb24d813314-node-bootstrap-token\") pod \"machine-config-server-h786z\" (UID: \"bb6ecea9-8eb0-4ff6-a898-5cb24d813314\") " pod="openshift-machine-config-operator/machine-config-server-h786z" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.223393 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/bb6ecea9-8eb0-4ff6-a898-5cb24d813314-certs\") pod \"machine-config-server-h786z\" (UID: \"bb6ecea9-8eb0-4ff6-a898-5cb24d813314\") " pod="openshift-machine-config-operator/machine-config-server-h786z" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.223434 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/25199d70-97f2-4513-a2e4-301c9ce3f85a-config-volume\") pod \"dns-default-fjz5z\" (UID: \"25199d70-97f2-4513-a2e4-301c9ce3f85a\") " pod="openshift-dns/dns-default-fjz5z" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.223494 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8d8x5\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.223556 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/25199d70-97f2-4513-a2e4-301c9ce3f85a-metrics-tls\") pod \"dns-default-fjz5z\" (UID: \"25199d70-97f2-4513-a2e4-301c9ce3f85a\") " pod="openshift-dns/dns-default-fjz5z" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.225149 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/25199d70-97f2-4513-a2e4-301c9ce3f85a-config-volume\") pod \"dns-default-fjz5z\" (UID: \"25199d70-97f2-4513-a2e4-301c9ce3f85a\") " pod="openshift-dns/dns-default-fjz5z" Oct 08 00:08:44 crc kubenswrapper[4967]: E1008 00:08:44.225998 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 00:08:44.725982937 +0000 UTC m=+123.376517586 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8d8x5" (UID: "1f06a5d1-8ff8-4d73-8939-0fac4252ea9c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.236266 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/bb6ecea9-8eb0-4ff6-a898-5cb24d813314-node-bootstrap-token\") pod \"machine-config-server-h786z\" (UID: \"bb6ecea9-8eb0-4ff6-a898-5cb24d813314\") " pod="openshift-machine-config-operator/machine-config-server-h786z" Oct 08 00:08:44 crc kubenswrapper[4967]: E1008 00:08:44.238030 4967 projected.go:288] Couldn't get configMap openshift-service-ca-operator/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.238465 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/bb6ecea9-8eb0-4ff6-a898-5cb24d813314-certs\") pod \"machine-config-server-h786z\" (UID: \"bb6ecea9-8eb0-4ff6-a898-5cb24d813314\") " pod="openshift-machine-config-operator/machine-config-server-h786z" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.243125 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.246069 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/25199d70-97f2-4513-a2e4-301c9ce3f85a-metrics-tls\") pod \"dns-default-fjz5z\" (UID: \"25199d70-97f2-4513-a2e4-301c9ce3f85a\") " pod="openshift-dns/dns-default-fjz5z" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.263844 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-tmj7c" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.263984 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ltttx" event={"ID":"f3ea161f-2b22-466c-b49b-9f60e9e3437a","Type":"ContainerStarted","Data":"035b2ebee2246e701b54099eec0f3cae187ec72b4b0f5a6c7c6a16d734971017"} Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.264002 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-pruner-29331360-494rn" event={"ID":"84cd9a7a-bb01-43de-9055-deccbc8529b2","Type":"ContainerStarted","Data":"8a3b4d3efcc161d3a54910bdf7e3aaf8f349acedd7bdf581c2b31f74eed0f3ae"} Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.264012 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-pruner-29331360-494rn" event={"ID":"84cd9a7a-bb01-43de-9055-deccbc8529b2","Type":"ContainerStarted","Data":"20c016534c059bad4d1f03867754c83e11b1d339f8550a6b47852de1dae5f9da"} Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.264021 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-p9jtz" event={"ID":"86941755-9852-4757-82b6-c4aa2c3ffa43","Type":"ContainerStarted","Data":"3b82bf379299a5974d40e7ff266e79263ffcd777b50bc98e63cbd2ec2e38bf30"} Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.264052 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-p9jtz" event={"ID":"86941755-9852-4757-82b6-c4aa2c3ffa43","Type":"ContainerStarted","Data":"a8d864775e93384ccf97ccc488ebf9446959f2e59cd053e4e2b875583b88b4dd"} Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.264064 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-b5mzq" event={"ID":"6801f4d5-5cce-4276-833a-20c81735e3be","Type":"ContainerStarted","Data":"50436656f893f9410ff323e1170578c38ca92d49f60920e4b5f74916907156aa"} Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.273701 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4kzvf" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.275428 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.283732 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 08 00:08:44 crc kubenswrapper[4967]: E1008 00:08:44.317900 4967 projected.go:288] Couldn't get configMap openshift-kube-storage-version-migrator/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.323585 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.324707 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.326490 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Oct 08 00:08:44 crc kubenswrapper[4967]: E1008 00:08:44.327592 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 00:08:44.827554016 +0000 UTC m=+123.478088665 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.328035 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c2302f69-8556-43c8-9451-145211206874-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-n69qz\" (UID: \"c2302f69-8556-43c8-9451-145211206874\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n69qz" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.334251 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c2302f69-8556-43c8-9451-145211206874-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-n69qz\" (UID: \"c2302f69-8556-43c8-9451-145211206874\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n69qz" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.360510 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.364201 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.364536 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-25pkg" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.365111 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2ghtr" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.385073 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.397958 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-tlvk7" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.402685 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Oct 08 00:08:44 crc kubenswrapper[4967]: E1008 00:08:44.420534 4967 projected.go:194] Error preparing data for projected volume kube-api-access-c7plb for pod openshift-etcd-operator/etcd-operator-b45778765-fqmxp: failed to sync configmap cache: timed out waiting for the condition Oct 08 00:08:44 crc kubenswrapper[4967]: E1008 00:08:44.420617 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/e873b76e-e526-4a38-a2d2-9b23ac147852-kube-api-access-c7plb podName:e873b76e-e526-4a38-a2d2-9b23ac147852 nodeName:}" failed. No retries permitted until 2025-10-08 00:08:44.920598143 +0000 UTC m=+123.571132792 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-c7plb" (UniqueName: "kubernetes.io/projected/e873b76e-e526-4a38-a2d2-9b23ac147852-kube-api-access-c7plb") pod "etcd-operator-b45778765-fqmxp" (UID: "e873b76e-e526-4a38-a2d2-9b23ac147852") : failed to sync configmap cache: timed out waiting for the condition Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.425790 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Oct 08 00:08:44 crc kubenswrapper[4967]: E1008 00:08:44.427151 4967 projected.go:194] Error preparing data for projected volume kube-api-access-vcvj4 for pod openshift-ingress-operator/ingress-operator-5b745b69d9-wp2tt: failed to sync configmap cache: timed out waiting for the condition Oct 08 00:08:44 crc kubenswrapper[4967]: E1008 00:08:44.427308 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4481eaf4-ecf4-46d3-be86-f028e4e33970-kube-api-access-vcvj4 podName:4481eaf4-ecf4-46d3-be86-f028e4e33970 nodeName:}" failed. No retries permitted until 2025-10-08 00:08:44.92727791 +0000 UTC m=+123.577812559 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-vcvj4" (UniqueName: "kubernetes.io/projected/4481eaf4-ecf4-46d3-be86-f028e4e33970-kube-api-access-vcvj4") pod "ingress-operator-5b745b69d9-wp2tt" (UID: "4481eaf4-ecf4-46d3-be86-f028e4e33970") : failed to sync configmap cache: timed out waiting for the condition Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.445524 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.448150 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8d8x5\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:08:44 crc kubenswrapper[4967]: E1008 00:08:44.450564 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 00:08:44.950543386 +0000 UTC m=+123.601078035 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8d8x5" (UID: "1f06a5d1-8ff8-4d73-8939-0fac4252ea9c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.484063 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.488822 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.520920 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rsxc6"] Oct 08 00:08:44 crc kubenswrapper[4967]: E1008 00:08:44.525889 4967 projected.go:194] Error preparing data for projected volume kube-api-access-br729 for pod hostpath-provisioner/csi-hostpathplugin-t65kf: failed to sync configmap cache: timed out waiting for the condition Oct 08 00:08:44 crc kubenswrapper[4967]: E1008 00:08:44.525973 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/aa6ab9dd-70bc-4834-a3be-f934e73c15c5-kube-api-access-br729 podName:aa6ab9dd-70bc-4834-a3be-f934e73c15c5 nodeName:}" failed. No retries permitted until 2025-10-08 00:08:45.025955957 +0000 UTC m=+123.676490606 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-br729" (UniqueName: "kubernetes.io/projected/aa6ab9dd-70bc-4834-a3be-f934e73c15c5-kube-api-access-br729") pod "csi-hostpathplugin-t65kf" (UID: "aa6ab9dd-70bc-4834-a3be-f934e73c15c5") : failed to sync configmap cache: timed out waiting for the condition Oct 08 00:08:44 crc kubenswrapper[4967]: E1008 00:08:44.529461 4967 projected.go:194] Error preparing data for projected volume kube-api-access-hkw6t for pod openshift-service-ca/service-ca-9c57cc56f-ghwrl: failed to sync configmap cache: timed out waiting for the condition Oct 08 00:08:44 crc kubenswrapper[4967]: E1008 00:08:44.529520 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/2c9d9d1d-3e68-47b5-b649-f193653c1922-kube-api-access-hkw6t podName:2c9d9d1d-3e68-47b5-b649-f193653c1922 nodeName:}" failed. No retries permitted until 2025-10-08 00:08:45.029505725 +0000 UTC m=+123.680040374 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-hkw6t" (UniqueName: "kubernetes.io/projected/2c9d9d1d-3e68-47b5-b649-f193653c1922-kube-api-access-hkw6t") pod "service-ca-9c57cc56f-ghwrl" (UID: "2c9d9d1d-3e68-47b5-b649-f193653c1922") : failed to sync configmap cache: timed out waiting for the condition Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.535788 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Oct 08 00:08:44 crc kubenswrapper[4967]: E1008 00:08:44.538718 4967 projected.go:194] Error preparing data for projected volume kube-api-access-d8rcj for pod openshift-ingress/router-default-5444994796-p9kxf: failed to sync configmap cache: timed out waiting for the condition Oct 08 00:08:44 crc kubenswrapper[4967]: E1008 00:08:44.538774 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/ee7dadb3-3ab7-4790-9b67-7f7a5bf7604f-kube-api-access-d8rcj podName:ee7dadb3-3ab7-4790-9b67-7f7a5bf7604f nodeName:}" failed. No retries permitted until 2025-10-08 00:08:45.038757574 +0000 UTC m=+123.689292223 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-d8rcj" (UniqueName: "kubernetes.io/projected/ee7dadb3-3ab7-4790-9b67-7f7a5bf7604f-kube-api-access-d8rcj") pod "router-default-5444994796-p9kxf" (UID: "ee7dadb3-3ab7-4790-9b67-7f7a5bf7604f") : failed to sync configmap cache: timed out waiting for the condition Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.552537 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 00:08:44 crc kubenswrapper[4967]: E1008 00:08:44.553001 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 00:08:45.052984907 +0000 UTC m=+123.703519556 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.556591 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.557328 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-lpxcq" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.562690 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Oct 08 00:08:44 crc kubenswrapper[4967]: E1008 00:08:44.573371 4967 projected.go:194] Error preparing data for projected volume kube-api-access-vtx28 for pod openshift-service-ca-operator/service-ca-operator-777779d784-kqmk8: failed to sync configmap cache: timed out waiting for the condition Oct 08 00:08:44 crc kubenswrapper[4967]: E1008 00:08:44.573449 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/e47422e6-8d98-4115-b59d-615218bfd4ae-kube-api-access-vtx28 podName:e47422e6-8d98-4115-b59d-615218bfd4ae nodeName:}" failed. No retries permitted until 2025-10-08 00:08:45.073427064 +0000 UTC m=+123.723961713 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-vtx28" (UniqueName: "kubernetes.io/projected/e47422e6-8d98-4115-b59d-615218bfd4ae-kube-api-access-vtx28") pod "service-ca-operator-777779d784-kqmk8" (UID: "e47422e6-8d98-4115-b59d-615218bfd4ae") : failed to sync configmap cache: timed out waiting for the condition Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.613089 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Oct 08 00:08:44 crc kubenswrapper[4967]: E1008 00:08:44.618228 4967 projected.go:194] Error preparing data for projected volume kube-api-access-9jvkr for pod openshift-kube-storage-version-migrator/migrator-59844c95c7-zkgkc: failed to sync configmap cache: timed out waiting for the condition Oct 08 00:08:44 crc kubenswrapper[4967]: E1008 00:08:44.618296 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9876428c-dc70-42b1-886b-2c8dfff20039-kube-api-access-9jvkr podName:9876428c-dc70-42b1-886b-2c8dfff20039 nodeName:}" failed. No retries permitted until 2025-10-08 00:08:45.118270427 +0000 UTC m=+123.768805076 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-9jvkr" (UniqueName: "kubernetes.io/projected/9876428c-dc70-42b1-886b-2c8dfff20039-kube-api-access-9jvkr") pod "migrator-59844c95c7-zkgkc" (UID: "9876428c-dc70-42b1-886b-2c8dfff20039") : failed to sync configmap cache: timed out waiting for the condition Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.619608 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.624382 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.637798 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-fjz5z" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.638937 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-86lvz" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.662149 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.678122 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fg895"] Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.678387 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-h786z" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.679415 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8d8x5\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.679538 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Oct 08 00:08:44 crc kubenswrapper[4967]: E1008 00:08:44.679930 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 00:08:45.179913425 +0000 UTC m=+123.830448074 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8d8x5" (UID: "1f06a5d1-8ff8-4d73-8939-0fac4252ea9c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.682254 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c4cg2"] Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.683941 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.697406 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-vxdxg" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.715442 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.743733 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-pjlzq" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.756991 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.757201 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.758415 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n69qz" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.759530 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vh779\" (UniqueName: \"kubernetes.io/projected/64202fa4-cb80-4028-aa03-d53231f2d8f4-kube-api-access-vh779\") pod \"ingress-canary-hpx62\" (UID: \"64202fa4-cb80-4028-aa03-d53231f2d8f4\") " pod="openshift-ingress-canary/ingress-canary-hpx62" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.790465 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 00:08:44 crc kubenswrapper[4967]: E1008 00:08:44.791001 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 00:08:45.29098538 +0000 UTC m=+123.941520029 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.796353 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w22ps\" (UniqueName: \"kubernetes.io/projected/f79e3965-8fdc-48ba-be6c-2866348f2458-kube-api-access-w22ps\") pod \"marketplace-operator-79b997595-npdfq\" (UID: \"f79e3965-8fdc-48ba-be6c-2866348f2458\") " pod="openshift-marketplace/marketplace-operator-79b997595-npdfq" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.829944 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cpnmt"] Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.841415 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.844159 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-npdfq" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.894076 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8d8x5\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:08:44 crc kubenswrapper[4967]: E1008 00:08:44.894599 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 00:08:45.394581849 +0000 UTC m=+124.045116498 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8d8x5" (UID: "1f06a5d1-8ff8-4d73-8939-0fac4252ea9c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.911000 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.911756 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-hpx62" Oct 08 00:08:44 crc kubenswrapper[4967]: W1008 00:08:44.943827 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5f033f1c_0cb3_4dcf_ba2b_79593657b841.slice/crio-6740526dfcadd5b22e193e60f3c2935e11f59b379ca94d6c050052de9c96fcc3 WatchSource:0}: Error finding container 6740526dfcadd5b22e193e60f3c2935e11f59b379ca94d6c050052de9c96fcc3: Status 404 returned error can't find the container with id 6740526dfcadd5b22e193e60f3c2935e11f59b379ca94d6c050052de9c96fcc3 Oct 08 00:08:44 crc kubenswrapper[4967]: W1008 00:08:44.983948 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0aec761a_1efe_4817_9bca_cf5961e28ac5.slice/crio-e089d7b28fadaab847ada19eb33d0fc9c27c5795edd394da4dfbe7412c59f968 WatchSource:0}: Error finding container e089d7b28fadaab847ada19eb33d0fc9c27c5795edd394da4dfbe7412c59f968: Status 404 returned error can't find the container with id e089d7b28fadaab847ada19eb33d0fc9c27c5795edd394da4dfbe7412c59f968 Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.994583 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.994837 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vcvj4\" (UniqueName: \"kubernetes.io/projected/4481eaf4-ecf4-46d3-be86-f028e4e33970-kube-api-access-vcvj4\") pod \"ingress-operator-5b745b69d9-wp2tt\" (UID: \"4481eaf4-ecf4-46d3-be86-f028e4e33970\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wp2tt" Oct 08 00:08:44 crc kubenswrapper[4967]: I1008 00:08:44.994900 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c7plb\" (UniqueName: \"kubernetes.io/projected/e873b76e-e526-4a38-a2d2-9b23ac147852-kube-api-access-c7plb\") pod \"etcd-operator-b45778765-fqmxp\" (UID: \"e873b76e-e526-4a38-a2d2-9b23ac147852\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fqmxp" Oct 08 00:08:44 crc kubenswrapper[4967]: E1008 00:08:44.996808 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 00:08:45.496782725 +0000 UTC m=+124.147317374 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:45 crc kubenswrapper[4967]: I1008 00:08:45.000731 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c7plb\" (UniqueName: \"kubernetes.io/projected/e873b76e-e526-4a38-a2d2-9b23ac147852-kube-api-access-c7plb\") pod \"etcd-operator-b45778765-fqmxp\" (UID: \"e873b76e-e526-4a38-a2d2-9b23ac147852\") " pod="openshift-etcd-operator/etcd-operator-b45778765-fqmxp" Oct 08 00:08:45 crc kubenswrapper[4967]: I1008 00:08:45.003232 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v4kxw" podStartSLOduration=101.003214644 podStartE2EDuration="1m41.003214644s" podCreationTimestamp="2025-10-08 00:07:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:08:44.958644248 +0000 UTC m=+123.609178897" watchObservedRunningTime="2025-10-08 00:08:45.003214644 +0000 UTC m=+123.653749293" Oct 08 00:08:45 crc kubenswrapper[4967]: I1008 00:08:45.004244 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vcvj4\" (UniqueName: \"kubernetes.io/projected/4481eaf4-ecf4-46d3-be86-f028e4e33970-kube-api-access-vcvj4\") pod \"ingress-operator-5b745b69d9-wp2tt\" (UID: \"4481eaf4-ecf4-46d3-be86-f028e4e33970\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wp2tt" Oct 08 00:08:45 crc kubenswrapper[4967]: I1008 00:08:45.099076 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hkw6t\" (UniqueName: \"kubernetes.io/projected/2c9d9d1d-3e68-47b5-b649-f193653c1922-kube-api-access-hkw6t\") pod \"service-ca-9c57cc56f-ghwrl\" (UID: \"2c9d9d1d-3e68-47b5-b649-f193653c1922\") " pod="openshift-service-ca/service-ca-9c57cc56f-ghwrl" Oct 08 00:08:45 crc kubenswrapper[4967]: I1008 00:08:45.099113 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-br729\" (UniqueName: \"kubernetes.io/projected/aa6ab9dd-70bc-4834-a3be-f934e73c15c5-kube-api-access-br729\") pod \"csi-hostpathplugin-t65kf\" (UID: \"aa6ab9dd-70bc-4834-a3be-f934e73c15c5\") " pod="hostpath-provisioner/csi-hostpathplugin-t65kf" Oct 08 00:08:45 crc kubenswrapper[4967]: I1008 00:08:45.099158 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vtx28\" (UniqueName: \"kubernetes.io/projected/e47422e6-8d98-4115-b59d-615218bfd4ae-kube-api-access-vtx28\") pod \"service-ca-operator-777779d784-kqmk8\" (UID: \"e47422e6-8d98-4115-b59d-615218bfd4ae\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-kqmk8" Oct 08 00:08:45 crc kubenswrapper[4967]: I1008 00:08:45.099182 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d8rcj\" (UniqueName: \"kubernetes.io/projected/ee7dadb3-3ab7-4790-9b67-7f7a5bf7604f-kube-api-access-d8rcj\") pod \"router-default-5444994796-p9kxf\" (UID: \"ee7dadb3-3ab7-4790-9b67-7f7a5bf7604f\") " pod="openshift-ingress/router-default-5444994796-p9kxf" Oct 08 00:08:45 crc kubenswrapper[4967]: I1008 00:08:45.099249 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8d8x5\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:08:45 crc kubenswrapper[4967]: E1008 00:08:45.099544 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 00:08:45.599532163 +0000 UTC m=+124.250066812 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8d8x5" (UID: "1f06a5d1-8ff8-4d73-8939-0fac4252ea9c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:45 crc kubenswrapper[4967]: I1008 00:08:45.110515 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hkw6t\" (UniqueName: \"kubernetes.io/projected/2c9d9d1d-3e68-47b5-b649-f193653c1922-kube-api-access-hkw6t\") pod \"service-ca-9c57cc56f-ghwrl\" (UID: \"2c9d9d1d-3e68-47b5-b649-f193653c1922\") " pod="openshift-service-ca/service-ca-9c57cc56f-ghwrl" Oct 08 00:08:45 crc kubenswrapper[4967]: I1008 00:08:45.115663 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d8rcj\" (UniqueName: \"kubernetes.io/projected/ee7dadb3-3ab7-4790-9b67-7f7a5bf7604f-kube-api-access-d8rcj\") pod \"router-default-5444994796-p9kxf\" (UID: \"ee7dadb3-3ab7-4790-9b67-7f7a5bf7604f\") " pod="openshift-ingress/router-default-5444994796-p9kxf" Oct 08 00:08:45 crc kubenswrapper[4967]: I1008 00:08:45.118647 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-tlvk7" podStartSLOduration=102.118630186 podStartE2EDuration="1m42.118630186s" podCreationTimestamp="2025-10-08 00:07:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:08:45.075711932 +0000 UTC m=+123.726246581" watchObservedRunningTime="2025-10-08 00:08:45.118630186 +0000 UTC m=+123.769164835" Oct 08 00:08:45 crc kubenswrapper[4967]: I1008 00:08:45.126663 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-br729\" (UniqueName: \"kubernetes.io/projected/aa6ab9dd-70bc-4834-a3be-f934e73c15c5-kube-api-access-br729\") pod \"csi-hostpathplugin-t65kf\" (UID: \"aa6ab9dd-70bc-4834-a3be-f934e73c15c5\") " pod="hostpath-provisioner/csi-hostpathplugin-t65kf" Oct 08 00:08:45 crc kubenswrapper[4967]: I1008 00:08:45.127333 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vtx28\" (UniqueName: \"kubernetes.io/projected/e47422e6-8d98-4115-b59d-615218bfd4ae-kube-api-access-vtx28\") pod \"service-ca-operator-777779d784-kqmk8\" (UID: \"e47422e6-8d98-4115-b59d-615218bfd4ae\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-kqmk8" Oct 08 00:08:45 crc kubenswrapper[4967]: I1008 00:08:45.127722 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Oct 08 00:08:45 crc kubenswrapper[4967]: I1008 00:08:45.140289 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-fqmxp" Oct 08 00:08:45 crc kubenswrapper[4967]: I1008 00:08:45.169962 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-h786z" event={"ID":"bb6ecea9-8eb0-4ff6-a898-5cb24d813314","Type":"ContainerStarted","Data":"9b2c9e99e9f3df592c340c34bac2e5205a648d46b9cc3767327cf9e0aeca04cc"} Oct 08 00:08:45 crc kubenswrapper[4967]: I1008 00:08:45.182696 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-tzhwk" event={"ID":"0d5e578e-60b8-400c-9545-110600602c3c","Type":"ContainerStarted","Data":"2a575232d3ea66e0fd4e9c9dd645fac41a5a114b0f47a7b68c8fb5512f9cd008"} Oct 08 00:08:45 crc kubenswrapper[4967]: I1008 00:08:45.184032 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ltttx" event={"ID":"f3ea161f-2b22-466c-b49b-9f60e9e3437a","Type":"ContainerStarted","Data":"2fa42bd17216a48bb9a6a974fd6391cf5b19f54655db1395ff50a2fd5407c09b"} Oct 08 00:08:45 crc kubenswrapper[4967]: I1008 00:08:45.185490 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cpnmt" event={"ID":"0aec761a-1efe-4817-9bca-cf5961e28ac5","Type":"ContainerStarted","Data":"e089d7b28fadaab847ada19eb33d0fc9c27c5795edd394da4dfbe7412c59f968"} Oct 08 00:08:45 crc kubenswrapper[4967]: I1008 00:08:45.186523 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c4cg2" event={"ID":"5f033f1c-0cb3-4dcf-ba2b-79593657b841","Type":"ContainerStarted","Data":"6740526dfcadd5b22e193e60f3c2935e11f59b379ca94d6c050052de9c96fcc3"} Oct 08 00:08:45 crc kubenswrapper[4967]: I1008 00:08:45.200428 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 00:08:45 crc kubenswrapper[4967]: I1008 00:08:45.200568 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9jvkr\" (UniqueName: \"kubernetes.io/projected/9876428c-dc70-42b1-886b-2c8dfff20039-kube-api-access-9jvkr\") pod \"migrator-59844c95c7-zkgkc\" (UID: \"9876428c-dc70-42b1-886b-2c8dfff20039\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-zkgkc" Oct 08 00:08:45 crc kubenswrapper[4967]: E1008 00:08:45.200738 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 00:08:45.700717413 +0000 UTC m=+124.351252062 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:45 crc kubenswrapper[4967]: I1008 00:08:45.210450 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rjwhx" event={"ID":"bb075d42-7ef0-4ad4-b817-bdd611f0d7ed","Type":"ContainerStarted","Data":"e4093e6d3151d61d3c9d09caba0eb926dbaf43e111eb5d4f0502ff712a30385b"} Oct 08 00:08:45 crc kubenswrapper[4967]: I1008 00:08:45.216702 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-vkf46" podStartSLOduration=101.216656648 podStartE2EDuration="1m41.216656648s" podCreationTimestamp="2025-10-08 00:07:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:08:45.215401897 +0000 UTC m=+123.865936566" watchObservedRunningTime="2025-10-08 00:08:45.216656648 +0000 UTC m=+123.867191297" Oct 08 00:08:45 crc kubenswrapper[4967]: I1008 00:08:45.219482 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-ck4pd" event={"ID":"71b5b4fb-2e90-4b39-b976-6806ff38ed8e","Type":"ContainerStarted","Data":"e139c10f841eb5e127672382cbf2abaca1a1b781a6c356a0e379bb06428e352c"} Oct 08 00:08:45 crc kubenswrapper[4967]: I1008 00:08:45.221626 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rsxc6" event={"ID":"6a6c3149-2d76-48de-963a-c5ef798a569b","Type":"ContainerStarted","Data":"e22887e72fef0f819da40b375df68d5ef0e28828d1cf6b6c8853eb7dd195bec5"} Oct 08 00:08:45 crc kubenswrapper[4967]: I1008 00:08:45.227993 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Oct 08 00:08:45 crc kubenswrapper[4967]: I1008 00:08:45.228749 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9jvkr\" (UniqueName: \"kubernetes.io/projected/9876428c-dc70-42b1-886b-2c8dfff20039-kube-api-access-9jvkr\") pod \"migrator-59844c95c7-zkgkc\" (UID: \"9876428c-dc70-42b1-886b-2c8dfff20039\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-zkgkc" Oct 08 00:08:45 crc kubenswrapper[4967]: I1008 00:08:45.238728 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wp2tt" Oct 08 00:08:45 crc kubenswrapper[4967]: I1008 00:08:45.241994 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Oct 08 00:08:45 crc kubenswrapper[4967]: I1008 00:08:45.245852 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-kqmk8" Oct 08 00:08:45 crc kubenswrapper[4967]: I1008 00:08:45.287019 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Oct 08 00:08:45 crc kubenswrapper[4967]: I1008 00:08:45.297654 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-ghwrl" Oct 08 00:08:45 crc kubenswrapper[4967]: I1008 00:08:45.301853 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8d8x5\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:08:45 crc kubenswrapper[4967]: E1008 00:08:45.302357 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 00:08:45.802342813 +0000 UTC m=+124.452877462 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8d8x5" (UID: "1f06a5d1-8ff8-4d73-8939-0fac4252ea9c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:45 crc kubenswrapper[4967]: I1008 00:08:45.304202 4967 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Oct 08 00:08:45 crc kubenswrapper[4967]: I1008 00:08:45.308208 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-t65kf" Oct 08 00:08:45 crc kubenswrapper[4967]: I1008 00:08:45.382787 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Oct 08 00:08:45 crc kubenswrapper[4967]: I1008 00:08:45.389440 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-p9kxf" Oct 08 00:08:45 crc kubenswrapper[4967]: I1008 00:08:45.406119 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 00:08:45 crc kubenswrapper[4967]: E1008 00:08:45.407349 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 00:08:45.907326817 +0000 UTC m=+124.557861466 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:45 crc kubenswrapper[4967]: I1008 00:08:45.448352 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-zkgkc" Oct 08 00:08:45 crc kubenswrapper[4967]: I1008 00:08:45.436738 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Oct 08 00:08:45 crc kubenswrapper[4967]: I1008 00:08:45.508845 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8d8x5\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:08:45 crc kubenswrapper[4967]: E1008 00:08:45.509182 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 00:08:46.009171304 +0000 UTC m=+124.659705953 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8d8x5" (UID: "1f06a5d1-8ff8-4d73-8939-0fac4252ea9c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:45 crc kubenswrapper[4967]: I1008 00:08:45.521723 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5wd5r" podStartSLOduration=102.521704744 podStartE2EDuration="1m42.521704744s" podCreationTimestamp="2025-10-08 00:07:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:08:45.482790909 +0000 UTC m=+124.133325558" watchObservedRunningTime="2025-10-08 00:08:45.521704744 +0000 UTC m=+124.172239393" Oct 08 00:08:45 crc kubenswrapper[4967]: I1008 00:08:45.611098 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 00:08:45 crc kubenswrapper[4967]: E1008 00:08:45.611408 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 00:08:46.111388819 +0000 UTC m=+124.761923468 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:45 crc kubenswrapper[4967]: I1008 00:08:45.611722 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8d8x5\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:08:45 crc kubenswrapper[4967]: E1008 00:08:45.611981 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 00:08:46.111973653 +0000 UTC m=+124.762508302 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8d8x5" (UID: "1f06a5d1-8ff8-4d73-8939-0fac4252ea9c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:45 crc kubenswrapper[4967]: I1008 00:08:45.712119 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-wxsv2"] Oct 08 00:08:45 crc kubenswrapper[4967]: I1008 00:08:45.712538 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 00:08:45 crc kubenswrapper[4967]: E1008 00:08:45.712885 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 00:08:46.212859295 +0000 UTC m=+124.863393934 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:45 crc kubenswrapper[4967]: I1008 00:08:45.746122 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-b5mzq" podStartSLOduration=102.74610369 podStartE2EDuration="1m42.74610369s" podCreationTimestamp="2025-10-08 00:07:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:08:45.73841919 +0000 UTC m=+124.388953849" watchObservedRunningTime="2025-10-08 00:08:45.74610369 +0000 UTC m=+124.396638339" Oct 08 00:08:45 crc kubenswrapper[4967]: I1008 00:08:45.821984 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8d8x5\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:08:45 crc kubenswrapper[4967]: E1008 00:08:45.822430 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 00:08:46.322417473 +0000 UTC m=+124.972952122 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8d8x5" (UID: "1f06a5d1-8ff8-4d73-8939-0fac4252ea9c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:45 crc kubenswrapper[4967]: I1008 00:08:45.923066 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 00:08:45 crc kubenswrapper[4967]: E1008 00:08:45.923448 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 00:08:46.423428379 +0000 UTC m=+125.073963028 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:45 crc kubenswrapper[4967]: E1008 00:08:45.924286 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 00:08:46.424261849 +0000 UTC m=+125.074796498 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8d8x5" (UID: "1f06a5d1-8ff8-4d73-8939-0fac4252ea9c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:45 crc kubenswrapper[4967]: I1008 00:08:45.925506 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8d8x5\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:08:46 crc kubenswrapper[4967]: I1008 00:08:46.030615 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 00:08:46 crc kubenswrapper[4967]: E1008 00:08:46.031232 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 00:08:46.531211302 +0000 UTC m=+125.181745951 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:46 crc kubenswrapper[4967]: I1008 00:08:46.094389 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-tmj7c" podStartSLOduration=103.094368259 podStartE2EDuration="1m43.094368259s" podCreationTimestamp="2025-10-08 00:07:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:08:46.068682192 +0000 UTC m=+124.719216831" watchObservedRunningTime="2025-10-08 00:08:46.094368259 +0000 UTC m=+124.744902908" Oct 08 00:08:46 crc kubenswrapper[4967]: I1008 00:08:46.133498 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8d8x5\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:08:46 crc kubenswrapper[4967]: E1008 00:08:46.134086 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 00:08:46.634062443 +0000 UTC m=+125.284597092 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8d8x5" (UID: "1f06a5d1-8ff8-4d73-8939-0fac4252ea9c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:46 crc kubenswrapper[4967]: I1008 00:08:46.176162 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-jjhb5" podStartSLOduration=103.176135886 podStartE2EDuration="1m43.176135886s" podCreationTimestamp="2025-10-08 00:07:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:08:46.173240585 +0000 UTC m=+124.823775244" watchObservedRunningTime="2025-10-08 00:08:46.176135886 +0000 UTC m=+124.826670535" Oct 08 00:08:46 crc kubenswrapper[4967]: I1008 00:08:46.235302 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 00:08:46 crc kubenswrapper[4967]: E1008 00:08:46.235698 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 00:08:46.735682133 +0000 UTC m=+125.386216782 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:46 crc kubenswrapper[4967]: I1008 00:08:46.243339 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rsxc6" event={"ID":"6a6c3149-2d76-48de-963a-c5ef798a569b","Type":"ContainerStarted","Data":"013b58f959f21d338a39e5127671a5869a0ee20b2ae54adfcb6f7ab432e6d2f8"} Oct 08 00:08:46 crc kubenswrapper[4967]: I1008 00:08:46.244323 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rsxc6" Oct 08 00:08:46 crc kubenswrapper[4967]: I1008 00:08:46.247686 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-tzhwk" event={"ID":"0d5e578e-60b8-400c-9545-110600602c3c","Type":"ContainerStarted","Data":"c74918818a366eff841c9f46f36963a7c57fe108eb7d382bb0c19d5316076793"} Oct 08 00:08:46 crc kubenswrapper[4967]: I1008 00:08:46.250820 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cpnmt" event={"ID":"0aec761a-1efe-4817-9bca-cf5961e28ac5","Type":"ContainerStarted","Data":"1521f06b5fe601a9219199bfd96c00190bcf186737c24a3dc0de5d2eb02857be"} Oct 08 00:08:46 crc kubenswrapper[4967]: I1008 00:08:46.251447 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cpnmt" Oct 08 00:08:46 crc kubenswrapper[4967]: I1008 00:08:46.253713 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c4cg2" event={"ID":"5f033f1c-0cb3-4dcf-ba2b-79593657b841","Type":"ContainerStarted","Data":"3471ea9a8f417e2b3ff55954f1a83c2f0a75c4af36d1aeebeae98b8252815982"} Oct 08 00:08:46 crc kubenswrapper[4967]: I1008 00:08:46.255515 4967 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-rsxc6 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.25:5443/healthz\": dial tcp 10.217.0.25:5443: connect: connection refused" start-of-body= Oct 08 00:08:46 crc kubenswrapper[4967]: I1008 00:08:46.255554 4967 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rsxc6" podUID="6a6c3149-2d76-48de-963a-c5ef798a569b" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.25:5443/healthz\": dial tcp 10.217.0.25:5443: connect: connection refused" Oct 08 00:08:46 crc kubenswrapper[4967]: I1008 00:08:46.256055 4967 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-cpnmt container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.24:8443/healthz\": dial tcp 10.217.0.24:8443: connect: connection refused" start-of-body= Oct 08 00:08:46 crc kubenswrapper[4967]: I1008 00:08:46.256082 4967 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cpnmt" podUID="0aec761a-1efe-4817-9bca-cf5961e28ac5" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.24:8443/healthz\": dial tcp 10.217.0.24:8443: connect: connection refused" Oct 08 00:08:46 crc kubenswrapper[4967]: I1008 00:08:46.263849 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-ck4pd" event={"ID":"71b5b4fb-2e90-4b39-b976-6806ff38ed8e","Type":"ContainerStarted","Data":"16329ae66edda8c377877fbc159fa316e4e887051e942e7c0cc411f306a8e6a9"} Oct 08 00:08:46 crc kubenswrapper[4967]: I1008 00:08:46.273810 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-p9kxf" event={"ID":"ee7dadb3-3ab7-4790-9b67-7f7a5bf7604f","Type":"ContainerStarted","Data":"48104fdeb593c9ec3776c5385dc662b247e7f8a58bd3c525196426a47d943245"} Oct 08 00:08:46 crc kubenswrapper[4967]: I1008 00:08:46.273887 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-p9kxf" event={"ID":"ee7dadb3-3ab7-4790-9b67-7f7a5bf7604f","Type":"ContainerStarted","Data":"471f69eb4f2113d603bd585f98ff8e2132aa9561efd2ef43212664f54d480cd2"} Oct 08 00:08:46 crc kubenswrapper[4967]: I1008 00:08:46.304180 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-wxsv2" event={"ID":"3164dad9-48a2-463c-a2c7-c668eec1ac77","Type":"ContainerStarted","Data":"6a53ee7487a287cdf569f2379c3dc0d396b475d5cbc0d9fbfe58cfa0cb83f210"} Oct 08 00:08:46 crc kubenswrapper[4967]: I1008 00:08:46.304226 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-wxsv2" event={"ID":"3164dad9-48a2-463c-a2c7-c668eec1ac77","Type":"ContainerStarted","Data":"e5375da06eee6812e3d6bc0f792f707e9df647bf3fff193765cd59dffb6489a4"} Oct 08 00:08:46 crc kubenswrapper[4967]: I1008 00:08:46.327056 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fg895" event={"ID":"af03bd84-939d-4a5e-9336-7f53108cdc33","Type":"ContainerStarted","Data":"ff5ad96b39c6ec4c7f269d177bf9b5a607d652dc9ead2e46f9242a26ff53c9fb"} Oct 08 00:08:46 crc kubenswrapper[4967]: I1008 00:08:46.327095 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fg895" event={"ID":"af03bd84-939d-4a5e-9336-7f53108cdc33","Type":"ContainerStarted","Data":"f368224f3c9f80743c1e739ab00905414b96f02be4cde7dcbb417e5f785db668"} Oct 08 00:08:46 crc kubenswrapper[4967]: I1008 00:08:46.339399 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8d8x5\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:08:46 crc kubenswrapper[4967]: E1008 00:08:46.344973 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 00:08:46.844959334 +0000 UTC m=+125.495493973 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8d8x5" (UID: "1f06a5d1-8ff8-4d73-8939-0fac4252ea9c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:46 crc kubenswrapper[4967]: I1008 00:08:46.364658 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-h786z" event={"ID":"bb6ecea9-8eb0-4ff6-a898-5cb24d813314","Type":"ContainerStarted","Data":"67c6e0b3b93415de2d3e487790bb50a638b03ee60fdb11e45bcce14169771c4b"} Oct 08 00:08:46 crc kubenswrapper[4967]: I1008 00:08:46.389415 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-25pkg"] Oct 08 00:08:46 crc kubenswrapper[4967]: I1008 00:08:46.399997 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-p9kxf" Oct 08 00:08:46 crc kubenswrapper[4967]: I1008 00:08:46.410338 4967 patch_prober.go:28] interesting pod/router-default-5444994796-p9kxf container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Oct 08 00:08:46 crc kubenswrapper[4967]: I1008 00:08:46.410386 4967 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-p9kxf" podUID="ee7dadb3-3ab7-4790-9b67-7f7a5bf7604f" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Oct 08 00:08:46 crc kubenswrapper[4967]: I1008 00:08:46.420306 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4kzvf"] Oct 08 00:08:46 crc kubenswrapper[4967]: I1008 00:08:46.479078 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-pruner-29331360-494rn" podStartSLOduration=103.47905218 podStartE2EDuration="1m43.47905218s" podCreationTimestamp="2025-10-08 00:07:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:08:46.425816489 +0000 UTC m=+125.076351138" watchObservedRunningTime="2025-10-08 00:08:46.47905218 +0000 UTC m=+125.129586829" Oct 08 00:08:46 crc kubenswrapper[4967]: I1008 00:08:46.480427 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-ljjfc" podStartSLOduration=102.480419713 podStartE2EDuration="1m42.480419713s" podCreationTimestamp="2025-10-08 00:07:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:08:46.361490614 +0000 UTC m=+125.012025263" watchObservedRunningTime="2025-10-08 00:08:46.480419713 +0000 UTC m=+125.130954362" Oct 08 00:08:46 crc kubenswrapper[4967]: I1008 00:08:46.481823 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 00:08:46 crc kubenswrapper[4967]: E1008 00:08:46.483977 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 00:08:46.983954132 +0000 UTC m=+125.634488781 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:46 crc kubenswrapper[4967]: I1008 00:08:46.485818 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b5glr"] Oct 08 00:08:46 crc kubenswrapper[4967]: I1008 00:08:46.517697 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-vxdxg"] Oct 08 00:08:46 crc kubenswrapper[4967]: I1008 00:08:46.530858 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331360-t4jlf"] Oct 08 00:08:46 crc kubenswrapper[4967]: I1008 00:08:46.569549 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-tdqfv" podStartSLOduration=103.569529464 podStartE2EDuration="1m43.569529464s" podCreationTimestamp="2025-10-08 00:07:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:08:46.566790466 +0000 UTC m=+125.217325115" watchObservedRunningTime="2025-10-08 00:08:46.569529464 +0000 UTC m=+125.220064113" Oct 08 00:08:46 crc kubenswrapper[4967]: I1008 00:08:46.584427 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8d8x5\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:08:46 crc kubenswrapper[4967]: E1008 00:08:46.584810 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 00:08:47.084797423 +0000 UTC m=+125.735332072 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8d8x5" (UID: "1f06a5d1-8ff8-4d73-8939-0fac4252ea9c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:46 crc kubenswrapper[4967]: I1008 00:08:46.626863 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n69qz"] Oct 08 00:08:46 crc kubenswrapper[4967]: I1008 00:08:46.650205 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-fjz5z"] Oct 08 00:08:46 crc kubenswrapper[4967]: I1008 00:08:46.660961 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-pjlzq"] Oct 08 00:08:46 crc kubenswrapper[4967]: I1008 00:08:46.672697 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-2ghtr"] Oct 08 00:08:46 crc kubenswrapper[4967]: I1008 00:08:46.675209 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-86lvz"] Oct 08 00:08:46 crc kubenswrapper[4967]: I1008 00:08:46.679352 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-krqks" podStartSLOduration=102.679333497 podStartE2EDuration="1m42.679333497s" podCreationTimestamp="2025-10-08 00:07:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:08:46.673121623 +0000 UTC m=+125.323656272" watchObservedRunningTime="2025-10-08 00:08:46.679333497 +0000 UTC m=+125.329868146" Oct 08 00:08:46 crc kubenswrapper[4967]: I1008 00:08:46.693603 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-lpxcq"] Oct 08 00:08:46 crc kubenswrapper[4967]: I1008 00:08:46.699194 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 00:08:46 crc kubenswrapper[4967]: E1008 00:08:46.699549 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 00:08:47.199532629 +0000 UTC m=+125.850067278 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:46 crc kubenswrapper[4967]: W1008 00:08:46.763083 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcd33606b_2d0f_4f4b_bc86_b85ad4f56f83.slice/crio-d762c9a330b15c409a9243121497fffc944b1c9d3b1e3bd32552b1d586ac5be4 WatchSource:0}: Error finding container d762c9a330b15c409a9243121497fffc944b1c9d3b1e3bd32552b1d586ac5be4: Status 404 returned error can't find the container with id d762c9a330b15c409a9243121497fffc944b1c9d3b1e3bd32552b1d586ac5be4 Oct 08 00:08:46 crc kubenswrapper[4967]: I1008 00:08:46.801025 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8d8x5\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:08:46 crc kubenswrapper[4967]: E1008 00:08:46.801404 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 00:08:47.301392075 +0000 UTC m=+125.951926724 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8d8x5" (UID: "1f06a5d1-8ff8-4d73-8939-0fac4252ea9c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:46 crc kubenswrapper[4967]: I1008 00:08:46.828724 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-p9jtz" podStartSLOduration=103.828706423 podStartE2EDuration="1m43.828706423s" podCreationTimestamp="2025-10-08 00:07:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:08:46.789969252 +0000 UTC m=+125.440503901" watchObservedRunningTime="2025-10-08 00:08:46.828706423 +0000 UTC m=+125.479241072" Oct 08 00:08:46 crc kubenswrapper[4967]: I1008 00:08:46.861358 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-zkgkc"] Oct 08 00:08:46 crc kubenswrapper[4967]: I1008 00:08:46.863004 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-kqmk8"] Oct 08 00:08:46 crc kubenswrapper[4967]: I1008 00:08:46.892530 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-npdfq"] Oct 08 00:08:46 crc kubenswrapper[4967]: I1008 00:08:46.916008 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 00:08:46 crc kubenswrapper[4967]: E1008 00:08:46.916993 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 00:08:47.416977521 +0000 UTC m=+126.067512170 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:46 crc kubenswrapper[4967]: I1008 00:08:46.941413 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-hpx62"] Oct 08 00:08:46 crc kubenswrapper[4967]: I1008 00:08:46.961260 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-ghwrl"] Oct 08 00:08:46 crc kubenswrapper[4967]: I1008 00:08:46.988289 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-t65kf"] Oct 08 00:08:47 crc kubenswrapper[4967]: I1008 00:08:47.044515 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-fqmxp"] Oct 08 00:08:47 crc kubenswrapper[4967]: I1008 00:08:47.046631 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8d8x5\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:08:47 crc kubenswrapper[4967]: E1008 00:08:47.046937 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 00:08:47.546925265 +0000 UTC m=+126.197459914 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8d8x5" (UID: "1f06a5d1-8ff8-4d73-8939-0fac4252ea9c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:47 crc kubenswrapper[4967]: I1008 00:08:47.147985 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-wp2tt"] Oct 08 00:08:47 crc kubenswrapper[4967]: I1008 00:08:47.149358 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 00:08:47 crc kubenswrapper[4967]: E1008 00:08:47.149752 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 00:08:47.649737464 +0000 UTC m=+126.300272113 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:47 crc kubenswrapper[4967]: W1008 00:08:47.158997 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode873b76e_e526_4a38_a2d2_9b23ac147852.slice/crio-473e74bf9d33bb7f94a891c9d29985a54414f8e79199abf6fffa6cd158b75af7 WatchSource:0}: Error finding container 473e74bf9d33bb7f94a891c9d29985a54414f8e79199abf6fffa6cd158b75af7: Status 404 returned error can't find the container with id 473e74bf9d33bb7f94a891c9d29985a54414f8e79199abf6fffa6cd158b75af7 Oct 08 00:08:47 crc kubenswrapper[4967]: I1008 00:08:47.251380 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8d8x5\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:08:47 crc kubenswrapper[4967]: E1008 00:08:47.252914 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 00:08:47.752899593 +0000 UTC m=+126.403434242 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8d8x5" (UID: "1f06a5d1-8ff8-4d73-8939-0fac4252ea9c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:47 crc kubenswrapper[4967]: I1008 00:08:47.264917 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-ltttx" podStartSLOduration=103.264897202 podStartE2EDuration="1m43.264897202s" podCreationTimestamp="2025-10-08 00:07:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:08:47.264093991 +0000 UTC m=+125.914628630" watchObservedRunningTime="2025-10-08 00:08:47.264897202 +0000 UTC m=+125.915431851" Oct 08 00:08:47 crc kubenswrapper[4967]: I1008 00:08:47.357127 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 00:08:47 crc kubenswrapper[4967]: E1008 00:08:47.357650 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 00:08:47.857631521 +0000 UTC m=+126.508166170 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:47 crc kubenswrapper[4967]: I1008 00:08:47.397375 4967 patch_prober.go:28] interesting pod/router-default-5444994796-p9kxf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 00:08:47 crc kubenswrapper[4967]: [-]has-synced failed: reason withheld Oct 08 00:08:47 crc kubenswrapper[4967]: [+]process-running ok Oct 08 00:08:47 crc kubenswrapper[4967]: healthz check failed Oct 08 00:08:47 crc kubenswrapper[4967]: I1008 00:08:47.397431 4967 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-p9kxf" podUID="ee7dadb3-3ab7-4790-9b67-7f7a5bf7604f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 00:08:47 crc kubenswrapper[4967]: I1008 00:08:47.417602 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fg895" podStartSLOduration=104.417586448 podStartE2EDuration="1m44.417586448s" podCreationTimestamp="2025-10-08 00:07:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:08:47.415207609 +0000 UTC m=+126.065742258" watchObservedRunningTime="2025-10-08 00:08:47.417586448 +0000 UTC m=+126.068121097" Oct 08 00:08:47 crc kubenswrapper[4967]: I1008 00:08:47.420461 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wp2tt" event={"ID":"4481eaf4-ecf4-46d3-be86-f028e4e33970","Type":"ContainerStarted","Data":"a5fbbb0651d4c1cc30e5c42c9ca7a88c4d9c10246749fad277f916020294248e"} Oct 08 00:08:47 crc kubenswrapper[4967]: I1008 00:08:47.433412 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-ck4pd" Oct 08 00:08:47 crc kubenswrapper[4967]: I1008 00:08:47.433820 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-ck4pd" Oct 08 00:08:47 crc kubenswrapper[4967]: I1008 00:08:47.445350 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-fjz5z" event={"ID":"25199d70-97f2-4513-a2e4-301c9ce3f85a","Type":"ContainerStarted","Data":"9b9781d0ecdf5636d682ca37b5eaf20cfe3399b85e031bbf2e1228e3449a22db"} Oct 08 00:08:47 crc kubenswrapper[4967]: I1008 00:08:47.449109 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-vxdxg" event={"ID":"8e5ce93e-573e-456c-a08c-4737fb3c7a66","Type":"ContainerStarted","Data":"9ce6b4fd515045fd135e2687897ef01f9dd6b8134513ba4aeac41c5d147329a4"} Oct 08 00:08:47 crc kubenswrapper[4967]: I1008 00:08:47.449145 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-vxdxg" event={"ID":"8e5ce93e-573e-456c-a08c-4737fb3c7a66","Type":"ContainerStarted","Data":"35918649af6765dbdb832c9e032dae9485a317d42fe9ea4c2ad4f7861634bbd7"} Oct 08 00:08:47 crc kubenswrapper[4967]: I1008 00:08:47.458554 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8d8x5\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:08:47 crc kubenswrapper[4967]: E1008 00:08:47.458937 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 00:08:47.958925144 +0000 UTC m=+126.609459793 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8d8x5" (UID: "1f06a5d1-8ff8-4d73-8939-0fac4252ea9c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:47 crc kubenswrapper[4967]: I1008 00:08:47.467874 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-tzhwk" podStartSLOduration=103.467857235 podStartE2EDuration="1m43.467857235s" podCreationTimestamp="2025-10-08 00:07:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:08:47.466388018 +0000 UTC m=+126.116922677" watchObservedRunningTime="2025-10-08 00:08:47.467857235 +0000 UTC m=+126.118391884" Oct 08 00:08:47 crc kubenswrapper[4967]: I1008 00:08:47.477205 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331360-t4jlf" event={"ID":"bb260fdf-16d6-4771-b7ae-38a669ff2393","Type":"ContainerStarted","Data":"7782cfe4b15523fd5b3b4d5ad87a4a2bf3d12cfca08fc3be11370557a2830069"} Oct 08 00:08:47 crc kubenswrapper[4967]: I1008 00:08:47.477276 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331360-t4jlf" event={"ID":"bb260fdf-16d6-4771-b7ae-38a669ff2393","Type":"ContainerStarted","Data":"e72b6d516bfd9e6181cbf6d1914738858269573d66ed955b3111e061c3078155"} Oct 08 00:08:47 crc kubenswrapper[4967]: I1008 00:08:47.490515 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-t65kf" event={"ID":"aa6ab9dd-70bc-4834-a3be-f934e73c15c5","Type":"ContainerStarted","Data":"eb9bdd0ee9f8ca63cea0536e24acc39703a5c13d1f150da8f473f3b11ffb0d80"} Oct 08 00:08:47 crc kubenswrapper[4967]: I1008 00:08:47.493047 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rjwhx" podStartSLOduration=103.49303727 podStartE2EDuration="1m43.49303727s" podCreationTimestamp="2025-10-08 00:07:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:08:47.491470791 +0000 UTC m=+126.142005440" watchObservedRunningTime="2025-10-08 00:08:47.49303727 +0000 UTC m=+126.143571919" Oct 08 00:08:47 crc kubenswrapper[4967]: I1008 00:08:47.498028 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-fqmxp" event={"ID":"e873b76e-e526-4a38-a2d2-9b23ac147852","Type":"ContainerStarted","Data":"473e74bf9d33bb7f94a891c9d29985a54414f8e79199abf6fffa6cd158b75af7"} Oct 08 00:08:47 crc kubenswrapper[4967]: I1008 00:08:47.499830 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-npdfq" event={"ID":"f79e3965-8fdc-48ba-be6c-2866348f2458","Type":"ContainerStarted","Data":"5606f3698d1288dfa61fe48f4444791a2fcc5155f6c140d0adb486c975c97fa1"} Oct 08 00:08:47 crc kubenswrapper[4967]: I1008 00:08:47.501023 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b5glr" event={"ID":"e0d38c54-b3a4-46b1-af38-0f8c7d3dbf5f","Type":"ContainerStarted","Data":"71d1cf370c290f0300d6415d6a5d12e42a34ac85072bd04ec74e584e1c83425a"} Oct 08 00:08:47 crc kubenswrapper[4967]: I1008 00:08:47.501043 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b5glr" event={"ID":"e0d38c54-b3a4-46b1-af38-0f8c7d3dbf5f","Type":"ContainerStarted","Data":"b62493485e1a757238344b5f2376e4e369439b4e57e074bddf007b4aa75b355f"} Oct 08 00:08:47 crc kubenswrapper[4967]: I1008 00:08:47.503083 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4kzvf" event={"ID":"9fa40840-4cd4-41ea-9c97-8a76e3001eee","Type":"ContainerStarted","Data":"b12f14731888ebc30ca34ff8ac29ec7595e6f36b84eca97e29487a39c4c6b562"} Oct 08 00:08:47 crc kubenswrapper[4967]: I1008 00:08:47.503118 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4kzvf" event={"ID":"9fa40840-4cd4-41ea-9c97-8a76e3001eee","Type":"ContainerStarted","Data":"87583ddba02260c05e29b83a57ec8a66305732dd59419a4adc5f084224e53e17"} Oct 08 00:08:47 crc kubenswrapper[4967]: I1008 00:08:47.503935 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4kzvf" Oct 08 00:08:47 crc kubenswrapper[4967]: I1008 00:08:47.515881 4967 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-4kzvf container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.34:8443/healthz\": dial tcp 10.217.0.34:8443: connect: connection refused" start-of-body= Oct 08 00:08:47 crc kubenswrapper[4967]: I1008 00:08:47.515961 4967 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4kzvf" podUID="9fa40840-4cd4-41ea-9c97-8a76e3001eee" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.34:8443/healthz\": dial tcp 10.217.0.34:8443: connect: connection refused" Oct 08 00:08:47 crc kubenswrapper[4967]: I1008 00:08:47.516092 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-kqmk8" event={"ID":"e47422e6-8d98-4115-b59d-615218bfd4ae","Type":"ContainerStarted","Data":"b3e71da8fec16b9edfa5d1ace776c07e9c25477063b8d0a99069a8fa357eb062"} Oct 08 00:08:47 crc kubenswrapper[4967]: I1008 00:08:47.537037 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-hpx62" event={"ID":"64202fa4-cb80-4028-aa03-d53231f2d8f4","Type":"ContainerStarted","Data":"64070aa52c678bf52b54f310782f5c5423fe15868321ef387b9da02fa4659773"} Oct 08 00:08:47 crc kubenswrapper[4967]: I1008 00:08:47.546344 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-wxsv2" event={"ID":"3164dad9-48a2-463c-a2c7-c668eec1ac77","Type":"ContainerStarted","Data":"1192cbe149f81ad934be11cbdcde63d9ac0a26dad0ec0268ab05a715100800d6"} Oct 08 00:08:47 crc kubenswrapper[4967]: I1008 00:08:47.547626 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-zkgkc" event={"ID":"9876428c-dc70-42b1-886b-2c8dfff20039","Type":"ContainerStarted","Data":"618d5651e939d8b65f09ec6027191647f6f9e9a26e1e82d335132df343d8833c"} Oct 08 00:08:47 crc kubenswrapper[4967]: I1008 00:08:47.550118 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2ghtr" event={"ID":"c7b7dad0-eb2c-4c7b-ac09-48d58ed7b936","Type":"ContainerStarted","Data":"48f9457bdbf02d2ecd96a3f6bc987192750c70d783bbb399e9109a12deeb1cf3"} Oct 08 00:08:47 crc kubenswrapper[4967]: I1008 00:08:47.550152 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2ghtr" event={"ID":"c7b7dad0-eb2c-4c7b-ac09-48d58ed7b936","Type":"ContainerStarted","Data":"4347bc4338f4458b99831b1564079db20c3c9c451731b40f47130027f9354178"} Oct 08 00:08:47 crc kubenswrapper[4967]: I1008 00:08:47.560161 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rsxc6" podStartSLOduration=103.560135934 podStartE2EDuration="1m43.560135934s" podCreationTimestamp="2025-10-08 00:07:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:08:47.554167026 +0000 UTC m=+126.204701675" watchObservedRunningTime="2025-10-08 00:08:47.560135934 +0000 UTC m=+126.210670583" Oct 08 00:08:47 crc kubenswrapper[4967]: I1008 00:08:47.562200 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 00:08:47 crc kubenswrapper[4967]: E1008 00:08:47.563963 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 00:08:48.063949588 +0000 UTC m=+126.714484237 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:47 crc kubenswrapper[4967]: I1008 00:08:47.564962 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rjwhx" Oct 08 00:08:47 crc kubenswrapper[4967]: I1008 00:08:47.570653 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rjwhx" Oct 08 00:08:47 crc kubenswrapper[4967]: I1008 00:08:47.574596 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-25pkg" event={"ID":"b2693055-f9b3-4539-ac6d-2151aafe6c2e","Type":"ContainerStarted","Data":"8d94342afb8cac81c766f568487d52af720e56a31f50bb44b38ed33f98386a8b"} Oct 08 00:08:47 crc kubenswrapper[4967]: I1008 00:08:47.574653 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-25pkg" event={"ID":"b2693055-f9b3-4539-ac6d-2151aafe6c2e","Type":"ContainerStarted","Data":"072f61cc4745bd54c768672e2b115677997c2d408c38680af5e9690f4d5c3d28"} Oct 08 00:08:47 crc kubenswrapper[4967]: I1008 00:08:47.595377 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fg895" event={"ID":"af03bd84-939d-4a5e-9336-7f53108cdc33","Type":"ContainerStarted","Data":"a25b697de3fab0a40555e653678da4f28aed1beed3ce9d8fb255dfb9d6b99477"} Oct 08 00:08:47 crc kubenswrapper[4967]: I1008 00:08:47.595945 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rjwhx" Oct 08 00:08:47 crc kubenswrapper[4967]: I1008 00:08:47.601789 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-ghwrl" event={"ID":"2c9d9d1d-3e68-47b5-b649-f193653c1922","Type":"ContainerStarted","Data":"a4c6d0f254da2109c65447720f99bec9c22df6f297cd17efbcad4b6bcf620f4e"} Oct 08 00:08:47 crc kubenswrapper[4967]: I1008 00:08:47.614034 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-pjlzq" event={"ID":"7e86e7c5-1431-4606-bf22-7bc05205e8e8","Type":"ContainerStarted","Data":"11b73fa1c1dc548b25a436ca69e0ada49d3f20c61d45d903194fee964b7027f3"} Oct 08 00:08:47 crc kubenswrapper[4967]: I1008 00:08:47.614081 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-pjlzq" event={"ID":"7e86e7c5-1431-4606-bf22-7bc05205e8e8","Type":"ContainerStarted","Data":"5e90c523c3db3ef90a8dae783266903b260ab50b59a201ed341eb6a14b06392d"} Oct 08 00:08:47 crc kubenswrapper[4967]: I1008 00:08:47.615102 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cpnmt" podStartSLOduration=103.615088857 podStartE2EDuration="1m43.615088857s" podCreationTimestamp="2025-10-08 00:07:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:08:47.614265936 +0000 UTC m=+126.264800585" watchObservedRunningTime="2025-10-08 00:08:47.615088857 +0000 UTC m=+126.265623506" Oct 08 00:08:47 crc kubenswrapper[4967]: I1008 00:08:47.625544 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n69qz" event={"ID":"c2302f69-8556-43c8-9451-145211206874","Type":"ContainerStarted","Data":"c8af1d6a89b1987641397373f1f68786ad59aed0aa278d5686d270bd52ebdbbe"} Oct 08 00:08:47 crc kubenswrapper[4967]: I1008 00:08:47.643761 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-86lvz" event={"ID":"553037a9-2f32-4eee-8244-caa084e58566","Type":"ContainerStarted","Data":"080507bcb48fadbb8f2c32685504781fb14ce6addac7789e8d001789f82ae0eb"} Oct 08 00:08:47 crc kubenswrapper[4967]: I1008 00:08:47.676928 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8d8x5\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:08:47 crc kubenswrapper[4967]: E1008 00:08:47.678768 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 00:08:48.178750906 +0000 UTC m=+126.829285565 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8d8x5" (UID: "1f06a5d1-8ff8-4d73-8939-0fac4252ea9c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:47 crc kubenswrapper[4967]: I1008 00:08:47.695555 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-lpxcq" event={"ID":"cd33606b-2d0f-4f4b-bc86-b85ad4f56f83","Type":"ContainerStarted","Data":"579b172907e3b9378d1879309ff1efd3075f47589179f3f91d0fab738d629040"} Oct 08 00:08:47 crc kubenswrapper[4967]: I1008 00:08:47.695602 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-lpxcq" event={"ID":"cd33606b-2d0f-4f4b-bc86-b85ad4f56f83","Type":"ContainerStarted","Data":"d762c9a330b15c409a9243121497fffc944b1c9d3b1e3bd32552b1d586ac5be4"} Oct 08 00:08:47 crc kubenswrapper[4967]: I1008 00:08:47.704098 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rjwhx" Oct 08 00:08:47 crc kubenswrapper[4967]: I1008 00:08:47.704058 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-h786z" podStartSLOduration=8.704035443 podStartE2EDuration="8.704035443s" podCreationTimestamp="2025-10-08 00:08:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:08:47.664951413 +0000 UTC m=+126.315486062" watchObservedRunningTime="2025-10-08 00:08:47.704035443 +0000 UTC m=+126.354570092" Oct 08 00:08:47 crc kubenswrapper[4967]: I1008 00:08:47.714774 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cpnmt" Oct 08 00:08:47 crc kubenswrapper[4967]: I1008 00:08:47.736884 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-ck4pd" podStartSLOduration=104.736853877 podStartE2EDuration="1m44.736853877s" podCreationTimestamp="2025-10-08 00:07:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:08:47.703570801 +0000 UTC m=+126.354105470" watchObservedRunningTime="2025-10-08 00:08:47.736853877 +0000 UTC m=+126.387388526" Oct 08 00:08:47 crc kubenswrapper[4967]: I1008 00:08:47.747081 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-rsxc6" Oct 08 00:08:47 crc kubenswrapper[4967]: I1008 00:08:47.797058 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 00:08:47 crc kubenswrapper[4967]: E1008 00:08:47.799688 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 00:08:48.299648124 +0000 UTC m=+126.950182773 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:47 crc kubenswrapper[4967]: I1008 00:08:47.813532 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-c4cg2" podStartSLOduration=103.813512918 podStartE2EDuration="1m43.813512918s" podCreationTimestamp="2025-10-08 00:07:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:08:47.811502399 +0000 UTC m=+126.462037068" watchObservedRunningTime="2025-10-08 00:08:47.813512918 +0000 UTC m=+126.464047557" Oct 08 00:08:47 crc kubenswrapper[4967]: I1008 00:08:47.876980 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-p9kxf" podStartSLOduration=103.876963192 podStartE2EDuration="1m43.876963192s" podCreationTimestamp="2025-10-08 00:07:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:08:47.87489043 +0000 UTC m=+126.525425079" watchObservedRunningTime="2025-10-08 00:08:47.876963192 +0000 UTC m=+126.527497841" Oct 08 00:08:47 crc kubenswrapper[4967]: I1008 00:08:47.900404 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8d8x5\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:08:47 crc kubenswrapper[4967]: E1008 00:08:47.901349 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 00:08:48.400706481 +0000 UTC m=+127.051241130 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8d8x5" (UID: "1f06a5d1-8ff8-4d73-8939-0fac4252ea9c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:47 crc kubenswrapper[4967]: I1008 00:08:47.919799 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-wxsv2" podStartSLOduration=103.919783294 podStartE2EDuration="1m43.919783294s" podCreationTimestamp="2025-10-08 00:07:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:08:47.917463316 +0000 UTC m=+126.567997965" watchObservedRunningTime="2025-10-08 00:08:47.919783294 +0000 UTC m=+126.570317943" Oct 08 00:08:47 crc kubenswrapper[4967]: I1008 00:08:47.948142 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4kzvf" podStartSLOduration=103.948120977 podStartE2EDuration="1m43.948120977s" podCreationTimestamp="2025-10-08 00:07:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:08:47.944773054 +0000 UTC m=+126.595307703" watchObservedRunningTime="2025-10-08 00:08:47.948120977 +0000 UTC m=+126.598655626" Oct 08 00:08:47 crc kubenswrapper[4967]: I1008 00:08:47.954130 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-jjhb5" Oct 08 00:08:47 crc kubenswrapper[4967]: I1008 00:08:47.969294 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-lpxcq" podStartSLOduration=103.969273652 podStartE2EDuration="1m43.969273652s" podCreationTimestamp="2025-10-08 00:07:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:08:47.96800257 +0000 UTC m=+126.618537219" watchObservedRunningTime="2025-10-08 00:08:47.969273652 +0000 UTC m=+126.619808301" Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.010477 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 00:08:48 crc kubenswrapper[4967]: E1008 00:08:48.010822 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 00:08:48.510777081 +0000 UTC m=+127.161311730 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.011030 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8d8x5\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:08:48 crc kubenswrapper[4967]: E1008 00:08:48.011391 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 00:08:48.511375876 +0000 UTC m=+127.161910525 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8d8x5" (UID: "1f06a5d1-8ff8-4d73-8939-0fac4252ea9c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.035573 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-pjlzq" podStartSLOduration=104.035550086 podStartE2EDuration="1m44.035550086s" podCreationTimestamp="2025-10-08 00:07:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:08:48.01033002 +0000 UTC m=+126.660864669" watchObservedRunningTime="2025-10-08 00:08:48.035550086 +0000 UTC m=+126.686084735" Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.056262 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-vxdxg" podStartSLOduration=105.056245009 podStartE2EDuration="1m45.056245009s" podCreationTimestamp="2025-10-08 00:07:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:08:48.037868573 +0000 UTC m=+126.688403222" watchObservedRunningTime="2025-10-08 00:08:48.056245009 +0000 UTC m=+126.706779658" Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.085582 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-25pkg" podStartSLOduration=104.085567756 podStartE2EDuration="1m44.085567756s" podCreationTimestamp="2025-10-08 00:07:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:08:48.083887205 +0000 UTC m=+126.734421854" watchObservedRunningTime="2025-10-08 00:08:48.085567756 +0000 UTC m=+126.736102405" Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.114110 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 00:08:48 crc kubenswrapper[4967]: E1008 00:08:48.114448 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 00:08:48.614432982 +0000 UTC m=+127.264967631 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.157878 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29331360-t4jlf" podStartSLOduration=105.157862909 podStartE2EDuration="1m45.157862909s" podCreationTimestamp="2025-10-08 00:07:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:08:48.156932296 +0000 UTC m=+126.807466945" watchObservedRunningTime="2025-10-08 00:08:48.157862909 +0000 UTC m=+126.808397558" Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.215223 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8d8x5\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:08:48 crc kubenswrapper[4967]: E1008 00:08:48.215650 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 00:08:48.715633542 +0000 UTC m=+127.366168191 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8d8x5" (UID: "1f06a5d1-8ff8-4d73-8939-0fac4252ea9c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.315957 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 00:08:48 crc kubenswrapper[4967]: E1008 00:08:48.316084 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 00:08:48.816068614 +0000 UTC m=+127.466603263 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.316154 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8d8x5\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:08:48 crc kubenswrapper[4967]: E1008 00:08:48.316448 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 00:08:48.816440593 +0000 UTC m=+127.466975242 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8d8x5" (UID: "1f06a5d1-8ff8-4d73-8939-0fac4252ea9c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.395303 4967 patch_prober.go:28] interesting pod/router-default-5444994796-p9kxf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 00:08:48 crc kubenswrapper[4967]: [-]has-synced failed: reason withheld Oct 08 00:08:48 crc kubenswrapper[4967]: [+]process-running ok Oct 08 00:08:48 crc kubenswrapper[4967]: healthz check failed Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.395355 4967 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-p9kxf" podUID="ee7dadb3-3ab7-4790-9b67-7f7a5bf7604f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.418279 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 00:08:48 crc kubenswrapper[4967]: E1008 00:08:48.418443 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 00:08:48.918396371 +0000 UTC m=+127.568931020 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.419009 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8d8x5\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:08:48 crc kubenswrapper[4967]: E1008 00:08:48.419263 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 00:08:48.919253723 +0000 UTC m=+127.569788372 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8d8x5" (UID: "1f06a5d1-8ff8-4d73-8939-0fac4252ea9c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.442536 4967 patch_prober.go:28] interesting pod/apiserver-76f77b778f-ck4pd container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Oct 08 00:08:48 crc kubenswrapper[4967]: [+]log ok Oct 08 00:08:48 crc kubenswrapper[4967]: [+]etcd ok Oct 08 00:08:48 crc kubenswrapper[4967]: [+]poststarthook/start-apiserver-admission-initializer ok Oct 08 00:08:48 crc kubenswrapper[4967]: [+]poststarthook/generic-apiserver-start-informers ok Oct 08 00:08:48 crc kubenswrapper[4967]: [+]poststarthook/max-in-flight-filter ok Oct 08 00:08:48 crc kubenswrapper[4967]: [+]poststarthook/storage-object-count-tracker-hook ok Oct 08 00:08:48 crc kubenswrapper[4967]: [+]poststarthook/image.openshift.io-apiserver-caches ok Oct 08 00:08:48 crc kubenswrapper[4967]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Oct 08 00:08:48 crc kubenswrapper[4967]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Oct 08 00:08:48 crc kubenswrapper[4967]: [+]poststarthook/project.openshift.io-projectcache ok Oct 08 00:08:48 crc kubenswrapper[4967]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Oct 08 00:08:48 crc kubenswrapper[4967]: [+]poststarthook/openshift.io-startinformers ok Oct 08 00:08:48 crc kubenswrapper[4967]: [+]poststarthook/openshift.io-restmapperupdater ok Oct 08 00:08:48 crc kubenswrapper[4967]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Oct 08 00:08:48 crc kubenswrapper[4967]: livez check failed Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.442662 4967 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-ck4pd" podUID="71b5b4fb-2e90-4b39-b976-6806ff38ed8e" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.496878 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-x8d4p"] Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.497890 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-x8d4p" Oct 08 00:08:48 crc kubenswrapper[4967]: W1008 00:08:48.499801 4967 reflector.go:561] object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g": failed to list *v1.Secret: secrets "certified-operators-dockercfg-4rs5g" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-marketplace": no relationship found between node 'crc' and this object Oct 08 00:08:48 crc kubenswrapper[4967]: E1008 00:08:48.499848 4967 reflector.go:158] "Unhandled Error" err="object-\"openshift-marketplace\"/\"certified-operators-dockercfg-4rs5g\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"certified-operators-dockercfg-4rs5g\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-marketplace\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.509420 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-x8d4p"] Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.519629 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 00:08:48 crc kubenswrapper[4967]: E1008 00:08:48.519907 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 00:08:49.019864207 +0000 UTC m=+127.670398856 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.520079 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8d8x5\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:08:48 crc kubenswrapper[4967]: E1008 00:08:48.520395 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 00:08:49.020366871 +0000 UTC m=+127.670901520 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8d8x5" (UID: "1f06a5d1-8ff8-4d73-8939-0fac4252ea9c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.620969 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 00:08:48 crc kubenswrapper[4967]: E1008 00:08:48.621148 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 00:08:49.12112137 +0000 UTC m=+127.771656029 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.621291 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8d8x5\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.621348 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/32d2d299-2970-40e9-9116-74a074c80eb1-utilities\") pod \"certified-operators-x8d4p\" (UID: \"32d2d299-2970-40e9-9116-74a074c80eb1\") " pod="openshift-marketplace/certified-operators-x8d4p" Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.621479 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/32d2d299-2970-40e9-9116-74a074c80eb1-catalog-content\") pod \"certified-operators-x8d4p\" (UID: \"32d2d299-2970-40e9-9116-74a074c80eb1\") " pod="openshift-marketplace/certified-operators-x8d4p" Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.621533 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-775qd\" (UniqueName: \"kubernetes.io/projected/32d2d299-2970-40e9-9116-74a074c80eb1-kube-api-access-775qd\") pod \"certified-operators-x8d4p\" (UID: \"32d2d299-2970-40e9-9116-74a074c80eb1\") " pod="openshift-marketplace/certified-operators-x8d4p" Oct 08 00:08:48 crc kubenswrapper[4967]: E1008 00:08:48.622072 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 00:08:49.122062553 +0000 UTC m=+127.772597202 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8d8x5" (UID: "1f06a5d1-8ff8-4d73-8939-0fac4252ea9c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.684574 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-9glfc"] Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.686298 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9glfc" Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.688760 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.706087 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9glfc"] Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.721408 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-ghwrl" event={"ID":"2c9d9d1d-3e68-47b5-b649-f193653c1922","Type":"ContainerStarted","Data":"b02f46283b715618f4b9bc15a02f13cacbf4781984f80dcc45a2e38e72b1b19b"} Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.722238 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 00:08:48 crc kubenswrapper[4967]: E1008 00:08:48.722473 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 00:08:49.222458293 +0000 UTC m=+127.872992942 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.722533 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f174780f-4c64-427f-b527-91a3871c271d-utilities\") pod \"community-operators-9glfc\" (UID: \"f174780f-4c64-427f-b527-91a3871c271d\") " pod="openshift-marketplace/community-operators-9glfc" Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.722571 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/32d2d299-2970-40e9-9116-74a074c80eb1-catalog-content\") pod \"certified-operators-x8d4p\" (UID: \"32d2d299-2970-40e9-9116-74a074c80eb1\") " pod="openshift-marketplace/certified-operators-x8d4p" Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.722593 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-775qd\" (UniqueName: \"kubernetes.io/projected/32d2d299-2970-40e9-9116-74a074c80eb1-kube-api-access-775qd\") pod \"certified-operators-x8d4p\" (UID: \"32d2d299-2970-40e9-9116-74a074c80eb1\") " pod="openshift-marketplace/certified-operators-x8d4p" Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.722683 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-blhnw\" (UniqueName: \"kubernetes.io/projected/f174780f-4c64-427f-b527-91a3871c271d-kube-api-access-blhnw\") pod \"community-operators-9glfc\" (UID: \"f174780f-4c64-427f-b527-91a3871c271d\") " pod="openshift-marketplace/community-operators-9glfc" Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.722719 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8d8x5\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.722740 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/32d2d299-2970-40e9-9116-74a074c80eb1-utilities\") pod \"certified-operators-x8d4p\" (UID: \"32d2d299-2970-40e9-9116-74a074c80eb1\") " pod="openshift-marketplace/certified-operators-x8d4p" Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.722758 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f174780f-4c64-427f-b527-91a3871c271d-catalog-content\") pod \"community-operators-9glfc\" (UID: \"f174780f-4c64-427f-b527-91a3871c271d\") " pod="openshift-marketplace/community-operators-9glfc" Oct 08 00:08:48 crc kubenswrapper[4967]: E1008 00:08:48.723450 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 00:08:49.223432507 +0000 UTC m=+127.873967156 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8d8x5" (UID: "1f06a5d1-8ff8-4d73-8939-0fac4252ea9c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.723735 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/32d2d299-2970-40e9-9116-74a074c80eb1-catalog-content\") pod \"certified-operators-x8d4p\" (UID: \"32d2d299-2970-40e9-9116-74a074c80eb1\") " pod="openshift-marketplace/certified-operators-x8d4p" Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.724718 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/32d2d299-2970-40e9-9116-74a074c80eb1-utilities\") pod \"certified-operators-x8d4p\" (UID: \"32d2d299-2970-40e9-9116-74a074c80eb1\") " pod="openshift-marketplace/certified-operators-x8d4p" Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.746005 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wp2tt" event={"ID":"4481eaf4-ecf4-46d3-be86-f028e4e33970","Type":"ContainerStarted","Data":"bb1bb632a8deb905e384f0a32d20f27c8f0d7cf39e832b5e8ce742acf116501d"} Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.746062 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wp2tt" event={"ID":"4481eaf4-ecf4-46d3-be86-f028e4e33970","Type":"ContainerStarted","Data":"6036a5a52cac54509cdf02a02e34a42dc6d365516a60db745925a885a68ce7be"} Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.754966 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-hpx62" event={"ID":"64202fa4-cb80-4028-aa03-d53231f2d8f4","Type":"ContainerStarted","Data":"c5a109bd809b91ebe3244b4329f3cd32de53c2dd8833278580e3a739d5261303"} Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.756697 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-ghwrl" podStartSLOduration=104.756678102 podStartE2EDuration="1m44.756678102s" podCreationTimestamp="2025-10-08 00:07:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:08:48.756512548 +0000 UTC m=+127.407047217" watchObservedRunningTime="2025-10-08 00:08:48.756678102 +0000 UTC m=+127.407212751" Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.770629 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-775qd\" (UniqueName: \"kubernetes.io/projected/32d2d299-2970-40e9-9116-74a074c80eb1-kube-api-access-775qd\") pod \"certified-operators-x8d4p\" (UID: \"32d2d299-2970-40e9-9116-74a074c80eb1\") " pod="openshift-marketplace/certified-operators-x8d4p" Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.773042 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-86lvz" event={"ID":"553037a9-2f32-4eee-8244-caa084e58566","Type":"ContainerStarted","Data":"e7f93f1de970aef35125023001ea7300ceac79e7fb97ecc1650b7b642c93ecf7"} Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.774244 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-kqmk8" event={"ID":"e47422e6-8d98-4115-b59d-615218bfd4ae","Type":"ContainerStarted","Data":"c32942f79ce3f79966eae6f1ac8f8a28b22441d8d2b7b84aa71236518ffb3eb0"} Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.788174 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-fqmxp" event={"ID":"e873b76e-e526-4a38-a2d2-9b23ac147852","Type":"ContainerStarted","Data":"dd281d5f405d0c2a8d9b9d835d25f8575d94e86a752c7153d1a6eeb6f1baf196"} Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.802777 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-hpx62" podStartSLOduration=9.802755904 podStartE2EDuration="9.802755904s" podCreationTimestamp="2025-10-08 00:08:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:08:48.800368476 +0000 UTC m=+127.450903125" watchObservedRunningTime="2025-10-08 00:08:48.802755904 +0000 UTC m=+127.453290553" Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.804413 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-25pkg" event={"ID":"b2693055-f9b3-4539-ac6d-2151aafe6c2e","Type":"ContainerStarted","Data":"a5c955e1c4d8119d81114687270a424eb8f6bee3de80dd0b0712f3871928310b"} Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.824368 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-npdfq" event={"ID":"f79e3965-8fdc-48ba-be6c-2866348f2458","Type":"ContainerStarted","Data":"34577d89f30ca78a8888d3edc5d18f6559afb5a278eb11dd7da7f5d531cb81d2"} Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.824444 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-npdfq" Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.825105 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.825275 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f174780f-4c64-427f-b527-91a3871c271d-utilities\") pod \"community-operators-9glfc\" (UID: \"f174780f-4c64-427f-b527-91a3871c271d\") " pod="openshift-marketplace/community-operators-9glfc" Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.825530 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-blhnw\" (UniqueName: \"kubernetes.io/projected/f174780f-4c64-427f-b527-91a3871c271d-kube-api-access-blhnw\") pod \"community-operators-9glfc\" (UID: \"f174780f-4c64-427f-b527-91a3871c271d\") " pod="openshift-marketplace/community-operators-9glfc" Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.825633 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f174780f-4c64-427f-b527-91a3871c271d-catalog-content\") pod \"community-operators-9glfc\" (UID: \"f174780f-4c64-427f-b527-91a3871c271d\") " pod="openshift-marketplace/community-operators-9glfc" Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.826082 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f174780f-4c64-427f-b527-91a3871c271d-catalog-content\") pod \"community-operators-9glfc\" (UID: \"f174780f-4c64-427f-b527-91a3871c271d\") " pod="openshift-marketplace/community-operators-9glfc" Oct 08 00:08:48 crc kubenswrapper[4967]: E1008 00:08:48.827699 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 00:08:49.327647682 +0000 UTC m=+127.978182331 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.829306 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f174780f-4c64-427f-b527-91a3871c271d-utilities\") pod \"community-operators-9glfc\" (UID: \"f174780f-4c64-427f-b527-91a3871c271d\") " pod="openshift-marketplace/community-operators-9glfc" Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.829460 4967 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-npdfq container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.29:8080/healthz\": dial tcp 10.217.0.29:8080: connect: connection refused" start-of-body= Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.838469 4967 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-npdfq" podUID="f79e3965-8fdc-48ba-be6c-2866348f2458" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.29:8080/healthz\": dial tcp 10.217.0.29:8080: connect: connection refused" Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.848151 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wp2tt" podStartSLOduration=104.848137451 podStartE2EDuration="1m44.848137451s" podCreationTimestamp="2025-10-08 00:07:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:08:48.847232577 +0000 UTC m=+127.497767226" watchObservedRunningTime="2025-10-08 00:08:48.848137451 +0000 UTC m=+127.498672100" Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.850421 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b5glr" event={"ID":"e0d38c54-b3a4-46b1-af38-0f8c7d3dbf5f","Type":"ContainerStarted","Data":"f1fca0f82097f8c4e13208cd7111223cb5586d8c00c428eab4e10b26b33e5920"} Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.850629 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b5glr" Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.859627 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n69qz" event={"ID":"c2302f69-8556-43c8-9451-145211206874","Type":"ContainerStarted","Data":"23e396913f50cd6e523f2ef932d6f8a44d991d47224e635aafb0962661dbf708"} Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.875997 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-blhnw\" (UniqueName: \"kubernetes.io/projected/f174780f-4c64-427f-b527-91a3871c271d-kube-api-access-blhnw\") pod \"community-operators-9glfc\" (UID: \"f174780f-4c64-427f-b527-91a3871c271d\") " pod="openshift-marketplace/community-operators-9glfc" Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.886726 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-zkgkc" event={"ID":"9876428c-dc70-42b1-886b-2c8dfff20039","Type":"ContainerStarted","Data":"3427581c70ed1886bb9a0979674d7178f685ae27ed29c69422c9f2af08b2277c"} Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.886802 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-zkgkc" event={"ID":"9876428c-dc70-42b1-886b-2c8dfff20039","Type":"ContainerStarted","Data":"555cdcc9a347970bad85a0c7cbeb584b246a90af61ff1efccd008e21cf9fdd63"} Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.892591 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-9sp7f"] Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.893975 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9sp7f" Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.917216 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-npdfq" podStartSLOduration=104.917184213 podStartE2EDuration="1m44.917184213s" podCreationTimestamp="2025-10-08 00:07:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:08:48.900061528 +0000 UTC m=+127.550596197" watchObservedRunningTime="2025-10-08 00:08:48.917184213 +0000 UTC m=+127.567718862" Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.919025 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2ghtr" event={"ID":"c7b7dad0-eb2c-4c7b-ac09-48d58ed7b936","Type":"ContainerStarted","Data":"264441c4ce6063cad562d74ef5e6a4765c328b18960c414a5d08c057d41ea0d9"} Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.920835 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9sp7f"] Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.927403 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t46zc\" (UniqueName: \"kubernetes.io/projected/a0f7251c-fe97-4f8d-8121-82f7010170ef-kube-api-access-t46zc\") pod \"certified-operators-9sp7f\" (UID: \"a0f7251c-fe97-4f8d-8121-82f7010170ef\") " pod="openshift-marketplace/certified-operators-9sp7f" Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.927446 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8d8x5\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.927486 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0f7251c-fe97-4f8d-8121-82f7010170ef-catalog-content\") pod \"certified-operators-9sp7f\" (UID: \"a0f7251c-fe97-4f8d-8121-82f7010170ef\") " pod="openshift-marketplace/certified-operators-9sp7f" Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.927554 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0f7251c-fe97-4f8d-8121-82f7010170ef-utilities\") pod \"certified-operators-9sp7f\" (UID: \"a0f7251c-fe97-4f8d-8121-82f7010170ef\") " pod="openshift-marketplace/certified-operators-9sp7f" Oct 08 00:08:48 crc kubenswrapper[4967]: E1008 00:08:48.928832 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 00:08:49.428814381 +0000 UTC m=+128.079349030 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8d8x5" (UID: "1f06a5d1-8ff8-4d73-8939-0fac4252ea9c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.935781 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-kqmk8" podStartSLOduration=104.935759594 podStartE2EDuration="1m44.935759594s" podCreationTimestamp="2025-10-08 00:07:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:08:48.934171174 +0000 UTC m=+127.584705823" watchObservedRunningTime="2025-10-08 00:08:48.935759594 +0000 UTC m=+127.586294243" Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.940970 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-fjz5z" event={"ID":"25199d70-97f2-4513-a2e4-301c9ce3f85a","Type":"ContainerStarted","Data":"92ee99c3a9627d9b4c4eb100fa9daf2392a56d6aa4f8ecaea2c45834ef4d2c7f"} Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.941047 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-fjz5z" Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.941062 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-fjz5z" event={"ID":"25199d70-97f2-4513-a2e4-301c9ce3f85a","Type":"ContainerStarted","Data":"1c471df95001bff5ed6274ae65b855e2ab9a26e0bec2e223d836600c81b55d71"} Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.958104 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-4kzvf" Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.974487 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-fqmxp" podStartSLOduration=104.974459124 podStartE2EDuration="1m44.974459124s" podCreationTimestamp="2025-10-08 00:07:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:08:48.960161969 +0000 UTC m=+127.610696608" watchObservedRunningTime="2025-10-08 00:08:48.974459124 +0000 UTC m=+127.624993773" Oct 08 00:08:48 crc kubenswrapper[4967]: I1008 00:08:48.990757 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-zkgkc" podStartSLOduration=104.990736218 podStartE2EDuration="1m44.990736218s" podCreationTimestamp="2025-10-08 00:07:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:08:48.989090606 +0000 UTC m=+127.639625255" watchObservedRunningTime="2025-10-08 00:08:48.990736218 +0000 UTC m=+127.641270867" Oct 08 00:08:49 crc kubenswrapper[4967]: I1008 00:08:49.009215 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n69qz" podStartSLOduration=105.009174594 podStartE2EDuration="1m45.009174594s" podCreationTimestamp="2025-10-08 00:07:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:08:49.007994155 +0000 UTC m=+127.658528804" watchObservedRunningTime="2025-10-08 00:08:49.009174594 +0000 UTC m=+127.659709243" Oct 08 00:08:49 crc kubenswrapper[4967]: I1008 00:08:49.018060 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9glfc" Oct 08 00:08:49 crc kubenswrapper[4967]: I1008 00:08:49.029874 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 00:08:49 crc kubenswrapper[4967]: I1008 00:08:49.030186 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0f7251c-fe97-4f8d-8121-82f7010170ef-catalog-content\") pod \"certified-operators-9sp7f\" (UID: \"a0f7251c-fe97-4f8d-8121-82f7010170ef\") " pod="openshift-marketplace/certified-operators-9sp7f" Oct 08 00:08:49 crc kubenswrapper[4967]: I1008 00:08:49.030324 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0f7251c-fe97-4f8d-8121-82f7010170ef-utilities\") pod \"certified-operators-9sp7f\" (UID: \"a0f7251c-fe97-4f8d-8121-82f7010170ef\") " pod="openshift-marketplace/certified-operators-9sp7f" Oct 08 00:08:49 crc kubenswrapper[4967]: I1008 00:08:49.030759 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t46zc\" (UniqueName: \"kubernetes.io/projected/a0f7251c-fe97-4f8d-8121-82f7010170ef-kube-api-access-t46zc\") pod \"certified-operators-9sp7f\" (UID: \"a0f7251c-fe97-4f8d-8121-82f7010170ef\") " pod="openshift-marketplace/certified-operators-9sp7f" Oct 08 00:08:49 crc kubenswrapper[4967]: E1008 00:08:49.031140 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 00:08:49.531123959 +0000 UTC m=+128.181658608 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:49 crc kubenswrapper[4967]: I1008 00:08:49.033219 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0f7251c-fe97-4f8d-8121-82f7010170ef-catalog-content\") pod \"certified-operators-9sp7f\" (UID: \"a0f7251c-fe97-4f8d-8121-82f7010170ef\") " pod="openshift-marketplace/certified-operators-9sp7f" Oct 08 00:08:49 crc kubenswrapper[4967]: I1008 00:08:49.035774 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0f7251c-fe97-4f8d-8121-82f7010170ef-utilities\") pod \"certified-operators-9sp7f\" (UID: \"a0f7251c-fe97-4f8d-8121-82f7010170ef\") " pod="openshift-marketplace/certified-operators-9sp7f" Oct 08 00:08:49 crc kubenswrapper[4967]: I1008 00:08:49.084919 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t46zc\" (UniqueName: \"kubernetes.io/projected/a0f7251c-fe97-4f8d-8121-82f7010170ef-kube-api-access-t46zc\") pod \"certified-operators-9sp7f\" (UID: \"a0f7251c-fe97-4f8d-8121-82f7010170ef\") " pod="openshift-marketplace/certified-operators-9sp7f" Oct 08 00:08:49 crc kubenswrapper[4967]: I1008 00:08:49.095975 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-fjz5z" podStartSLOduration=10.095938996 podStartE2EDuration="10.095938996s" podCreationTimestamp="2025-10-08 00:08:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:08:49.072739621 +0000 UTC m=+127.723274280" watchObservedRunningTime="2025-10-08 00:08:49.095938996 +0000 UTC m=+127.746473645" Oct 08 00:08:49 crc kubenswrapper[4967]: I1008 00:08:49.127591 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-4hhcd"] Oct 08 00:08:49 crc kubenswrapper[4967]: I1008 00:08:49.127819 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-2ghtr" podStartSLOduration=105.127788507 podStartE2EDuration="1m45.127788507s" podCreationTimestamp="2025-10-08 00:07:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:08:49.09567479 +0000 UTC m=+127.746209439" watchObservedRunningTime="2025-10-08 00:08:49.127788507 +0000 UTC m=+127.778323156" Oct 08 00:08:49 crc kubenswrapper[4967]: I1008 00:08:49.129575 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4hhcd" Oct 08 00:08:49 crc kubenswrapper[4967]: I1008 00:08:49.132854 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8d8x5\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:08:49 crc kubenswrapper[4967]: E1008 00:08:49.133347 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 00:08:49.633331063 +0000 UTC m=+128.283865712 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8d8x5" (UID: "1f06a5d1-8ff8-4d73-8939-0fac4252ea9c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:49 crc kubenswrapper[4967]: I1008 00:08:49.133531 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4hhcd"] Oct 08 00:08:49 crc kubenswrapper[4967]: I1008 00:08:49.134471 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b5glr" podStartSLOduration=105.134462142 podStartE2EDuration="1m45.134462142s" podCreationTimestamp="2025-10-08 00:07:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:08:49.123588102 +0000 UTC m=+127.774122761" watchObservedRunningTime="2025-10-08 00:08:49.134462142 +0000 UTC m=+127.784996791" Oct 08 00:08:49 crc kubenswrapper[4967]: I1008 00:08:49.234343 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 00:08:49 crc kubenswrapper[4967]: E1008 00:08:49.234796 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 00:08:49.73477668 +0000 UTC m=+128.385311329 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:49 crc kubenswrapper[4967]: I1008 00:08:49.235072 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8d8x5\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:08:49 crc kubenswrapper[4967]: I1008 00:08:49.235127 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8912b04d-fd70-497c-aab6-7490fc6ddf40-catalog-content\") pod \"community-operators-4hhcd\" (UID: \"8912b04d-fd70-497c-aab6-7490fc6ddf40\") " pod="openshift-marketplace/community-operators-4hhcd" Oct 08 00:08:49 crc kubenswrapper[4967]: I1008 00:08:49.235147 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwcx7\" (UniqueName: \"kubernetes.io/projected/8912b04d-fd70-497c-aab6-7490fc6ddf40-kube-api-access-fwcx7\") pod \"community-operators-4hhcd\" (UID: \"8912b04d-fd70-497c-aab6-7490fc6ddf40\") " pod="openshift-marketplace/community-operators-4hhcd" Oct 08 00:08:49 crc kubenswrapper[4967]: I1008 00:08:49.235162 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8912b04d-fd70-497c-aab6-7490fc6ddf40-utilities\") pod \"community-operators-4hhcd\" (UID: \"8912b04d-fd70-497c-aab6-7490fc6ddf40\") " pod="openshift-marketplace/community-operators-4hhcd" Oct 08 00:08:49 crc kubenswrapper[4967]: E1008 00:08:49.235486 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 00:08:49.735472737 +0000 UTC m=+128.386007386 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8d8x5" (UID: "1f06a5d1-8ff8-4d73-8939-0fac4252ea9c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:49 crc kubenswrapper[4967]: I1008 00:08:49.337114 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 08 00:08:49 crc kubenswrapper[4967]: I1008 00:08:49.341299 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-x8d4p" Oct 08 00:08:49 crc kubenswrapper[4967]: I1008 00:08:49.342932 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 00:08:49 crc kubenswrapper[4967]: I1008 00:08:49.343228 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8912b04d-fd70-497c-aab6-7490fc6ddf40-catalog-content\") pod \"community-operators-4hhcd\" (UID: \"8912b04d-fd70-497c-aab6-7490fc6ddf40\") " pod="openshift-marketplace/community-operators-4hhcd" Oct 08 00:08:49 crc kubenswrapper[4967]: I1008 00:08:49.343258 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwcx7\" (UniqueName: \"kubernetes.io/projected/8912b04d-fd70-497c-aab6-7490fc6ddf40-kube-api-access-fwcx7\") pod \"community-operators-4hhcd\" (UID: \"8912b04d-fd70-497c-aab6-7490fc6ddf40\") " pod="openshift-marketplace/community-operators-4hhcd" Oct 08 00:08:49 crc kubenswrapper[4967]: I1008 00:08:49.343275 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8912b04d-fd70-497c-aab6-7490fc6ddf40-utilities\") pod \"community-operators-4hhcd\" (UID: \"8912b04d-fd70-497c-aab6-7490fc6ddf40\") " pod="openshift-marketplace/community-operators-4hhcd" Oct 08 00:08:49 crc kubenswrapper[4967]: E1008 00:08:49.343330 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 00:08:49.843306862 +0000 UTC m=+128.493841511 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:49 crc kubenswrapper[4967]: I1008 00:08:49.343728 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8912b04d-fd70-497c-aab6-7490fc6ddf40-utilities\") pod \"community-operators-4hhcd\" (UID: \"8912b04d-fd70-497c-aab6-7490fc6ddf40\") " pod="openshift-marketplace/community-operators-4hhcd" Oct 08 00:08:49 crc kubenswrapper[4967]: I1008 00:08:49.343898 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8912b04d-fd70-497c-aab6-7490fc6ddf40-catalog-content\") pod \"community-operators-4hhcd\" (UID: \"8912b04d-fd70-497c-aab6-7490fc6ddf40\") " pod="openshift-marketplace/community-operators-4hhcd" Oct 08 00:08:49 crc kubenswrapper[4967]: I1008 00:08:49.347794 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9sp7f" Oct 08 00:08:49 crc kubenswrapper[4967]: I1008 00:08:49.370725 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwcx7\" (UniqueName: \"kubernetes.io/projected/8912b04d-fd70-497c-aab6-7490fc6ddf40-kube-api-access-fwcx7\") pod \"community-operators-4hhcd\" (UID: \"8912b04d-fd70-497c-aab6-7490fc6ddf40\") " pod="openshift-marketplace/community-operators-4hhcd" Oct 08 00:08:49 crc kubenswrapper[4967]: I1008 00:08:49.398702 4967 patch_prober.go:28] interesting pod/router-default-5444994796-p9kxf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 00:08:49 crc kubenswrapper[4967]: [-]has-synced failed: reason withheld Oct 08 00:08:49 crc kubenswrapper[4967]: [+]process-running ok Oct 08 00:08:49 crc kubenswrapper[4967]: healthz check failed Oct 08 00:08:49 crc kubenswrapper[4967]: I1008 00:08:49.398759 4967 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-p9kxf" podUID="ee7dadb3-3ab7-4790-9b67-7f7a5bf7604f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 00:08:49 crc kubenswrapper[4967]: I1008 00:08:49.448787 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8d8x5\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:08:49 crc kubenswrapper[4967]: E1008 00:08:49.449101 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 00:08:49.949089896 +0000 UTC m=+128.599624545 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8d8x5" (UID: "1f06a5d1-8ff8-4d73-8939-0fac4252ea9c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:49 crc kubenswrapper[4967]: I1008 00:08:49.457884 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4hhcd" Oct 08 00:08:49 crc kubenswrapper[4967]: I1008 00:08:49.514410 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9glfc"] Oct 08 00:08:49 crc kubenswrapper[4967]: I1008 00:08:49.550073 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 00:08:49 crc kubenswrapper[4967]: E1008 00:08:49.550518 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 00:08:50.050503981 +0000 UTC m=+128.701038630 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:49 crc kubenswrapper[4967]: I1008 00:08:49.654157 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8d8x5\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:08:49 crc kubenswrapper[4967]: E1008 00:08:49.654598 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 00:08:50.154579453 +0000 UTC m=+128.805114102 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8d8x5" (UID: "1f06a5d1-8ff8-4d73-8939-0fac4252ea9c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:49 crc kubenswrapper[4967]: I1008 00:08:49.700649 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-x8d4p"] Oct 08 00:08:49 crc kubenswrapper[4967]: W1008 00:08:49.735807 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod32d2d299_2970_40e9_9116_74a074c80eb1.slice/crio-4d7207b9b44c47786d92cec91a19c4ddf87acaa341e7d59d18393732ed3ef1b0 WatchSource:0}: Error finding container 4d7207b9b44c47786d92cec91a19c4ddf87acaa341e7d59d18393732ed3ef1b0: Status 404 returned error can't find the container with id 4d7207b9b44c47786d92cec91a19c4ddf87acaa341e7d59d18393732ed3ef1b0 Oct 08 00:08:49 crc kubenswrapper[4967]: I1008 00:08:49.757295 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 00:08:49 crc kubenswrapper[4967]: E1008 00:08:49.757661 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 00:08:50.257642638 +0000 UTC m=+128.908177287 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:49 crc kubenswrapper[4967]: I1008 00:08:49.861874 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8d8x5\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:08:49 crc kubenswrapper[4967]: E1008 00:08:49.862552 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 00:08:50.362539911 +0000 UTC m=+129.013074560 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8d8x5" (UID: "1f06a5d1-8ff8-4d73-8939-0fac4252ea9c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:49 crc kubenswrapper[4967]: I1008 00:08:49.918009 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4hhcd"] Oct 08 00:08:49 crc kubenswrapper[4967]: I1008 00:08:49.965702 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 00:08:49 crc kubenswrapper[4967]: E1008 00:08:49.966133 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 00:08:50.466107669 +0000 UTC m=+129.116642318 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:49 crc kubenswrapper[4967]: I1008 00:08:49.987203 4967 generic.go:334] "Generic (PLEG): container finished" podID="f174780f-4c64-427f-b527-91a3871c271d" containerID="e242ae473cda65bdf80d16e7b1ae127906db7b7f307e0b71b636d1caa3f3818c" exitCode=0 Oct 08 00:08:49 crc kubenswrapper[4967]: I1008 00:08:49.988721 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9glfc" event={"ID":"f174780f-4c64-427f-b527-91a3871c271d","Type":"ContainerDied","Data":"e242ae473cda65bdf80d16e7b1ae127906db7b7f307e0b71b636d1caa3f3818c"} Oct 08 00:08:49 crc kubenswrapper[4967]: I1008 00:08:49.992517 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9glfc" event={"ID":"f174780f-4c64-427f-b527-91a3871c271d","Type":"ContainerStarted","Data":"2a89b55c2d7c99f1995c42994d2f74ab6d2ad5cbb2a4bda1d5c74f9636741158"} Oct 08 00:08:49 crc kubenswrapper[4967]: I1008 00:08:49.996948 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x8d4p" event={"ID":"32d2d299-2970-40e9-9116-74a074c80eb1","Type":"ContainerStarted","Data":"4d7207b9b44c47786d92cec91a19c4ddf87acaa341e7d59d18393732ed3ef1b0"} Oct 08 00:08:50 crc kubenswrapper[4967]: I1008 00:08:50.016619 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-t65kf" event={"ID":"aa6ab9dd-70bc-4834-a3be-f934e73c15c5","Type":"ContainerStarted","Data":"87d7aa986bc6263200587a5d6c67303835f5c01b87e189f0950bc17ee2dd88d6"} Oct 08 00:08:50 crc kubenswrapper[4967]: I1008 00:08:50.020435 4967 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 00:08:50 crc kubenswrapper[4967]: I1008 00:08:50.048992 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-86lvz" event={"ID":"553037a9-2f32-4eee-8244-caa084e58566","Type":"ContainerStarted","Data":"c994e8ccdb4782f525196f7aeb20f49b718c3daa4e31b7d7867b2af4c974bdfe"} Oct 08 00:08:50 crc kubenswrapper[4967]: I1008 00:08:50.055078 4967 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-npdfq container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.29:8080/healthz\": dial tcp 10.217.0.29:8080: connect: connection refused" start-of-body= Oct 08 00:08:50 crc kubenswrapper[4967]: I1008 00:08:50.055116 4967 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-npdfq" podUID="f79e3965-8fdc-48ba-be6c-2866348f2458" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.29:8080/healthz\": dial tcp 10.217.0.29:8080: connect: connection refused" Oct 08 00:08:50 crc kubenswrapper[4967]: I1008 00:08:50.066984 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8d8x5\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:08:50 crc kubenswrapper[4967]: E1008 00:08:50.067294 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 00:08:50.567281079 +0000 UTC m=+129.217815718 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8d8x5" (UID: "1f06a5d1-8ff8-4d73-8939-0fac4252ea9c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:50 crc kubenswrapper[4967]: I1008 00:08:50.078619 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-86lvz" podStartSLOduration=106.078574649 podStartE2EDuration="1m46.078574649s" podCreationTimestamp="2025-10-08 00:07:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:08:50.07216923 +0000 UTC m=+128.722703879" watchObservedRunningTime="2025-10-08 00:08:50.078574649 +0000 UTC m=+128.729109298" Oct 08 00:08:50 crc kubenswrapper[4967]: I1008 00:08:50.168644 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 00:08:50 crc kubenswrapper[4967]: E1008 00:08:50.170579 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 00:08:50.6705457 +0000 UTC m=+129.321080349 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:50 crc kubenswrapper[4967]: I1008 00:08:50.192694 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9sp7f"] Oct 08 00:08:50 crc kubenswrapper[4967]: I1008 00:08:50.273723 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8d8x5\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:08:50 crc kubenswrapper[4967]: E1008 00:08:50.274031 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 00:08:50.774019047 +0000 UTC m=+129.424553696 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8d8x5" (UID: "1f06a5d1-8ff8-4d73-8939-0fac4252ea9c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:50 crc kubenswrapper[4967]: I1008 00:08:50.374583 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 00:08:50 crc kubenswrapper[4967]: E1008 00:08:50.374815 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 00:08:50.874785016 +0000 UTC m=+129.525319665 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:50 crc kubenswrapper[4967]: I1008 00:08:50.375091 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8d8x5\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:08:50 crc kubenswrapper[4967]: E1008 00:08:50.375404 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 00:08:50.875393461 +0000 UTC m=+129.525928100 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8d8x5" (UID: "1f06a5d1-8ff8-4d73-8939-0fac4252ea9c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:50 crc kubenswrapper[4967]: I1008 00:08:50.399288 4967 patch_prober.go:28] interesting pod/router-default-5444994796-p9kxf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 00:08:50 crc kubenswrapper[4967]: [-]has-synced failed: reason withheld Oct 08 00:08:50 crc kubenswrapper[4967]: [+]process-running ok Oct 08 00:08:50 crc kubenswrapper[4967]: healthz check failed Oct 08 00:08:50 crc kubenswrapper[4967]: I1008 00:08:50.399731 4967 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-p9kxf" podUID="ee7dadb3-3ab7-4790-9b67-7f7a5bf7604f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 00:08:50 crc kubenswrapper[4967]: I1008 00:08:50.483784 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 00:08:50 crc kubenswrapper[4967]: E1008 00:08:50.484652 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 00:08:50.98462756 +0000 UTC m=+129.635162209 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:50 crc kubenswrapper[4967]: I1008 00:08:50.585968 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8d8x5\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:08:50 crc kubenswrapper[4967]: E1008 00:08:50.586343 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 00:08:51.086324452 +0000 UTC m=+129.736859101 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8d8x5" (UID: "1f06a5d1-8ff8-4d73-8939-0fac4252ea9c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:50 crc kubenswrapper[4967]: I1008 00:08:50.687605 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 00:08:50 crc kubenswrapper[4967]: I1008 00:08:50.687806 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-rzc76"] Oct 08 00:08:50 crc kubenswrapper[4967]: E1008 00:08:50.687892 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 00:08:51.1878551 +0000 UTC m=+129.838389749 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:50 crc kubenswrapper[4967]: I1008 00:08:50.688134 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8d8x5\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:08:50 crc kubenswrapper[4967]: E1008 00:08:50.688636 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 00:08:51.18861871 +0000 UTC m=+129.839153359 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8d8x5" (UID: "1f06a5d1-8ff8-4d73-8939-0fac4252ea9c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:50 crc kubenswrapper[4967]: I1008 00:08:50.688833 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rzc76" Oct 08 00:08:50 crc kubenswrapper[4967]: I1008 00:08:50.701715 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 08 00:08:50 crc kubenswrapper[4967]: I1008 00:08:50.711409 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rzc76"] Oct 08 00:08:50 crc kubenswrapper[4967]: I1008 00:08:50.790150 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 00:08:50 crc kubenswrapper[4967]: E1008 00:08:50.790471 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 00:08:51.290432045 +0000 UTC m=+129.940966704 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:50 crc kubenswrapper[4967]: I1008 00:08:50.790518 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f0340c31-098f-4936-bd5e-4ef17c48a9f1-utilities\") pod \"redhat-marketplace-rzc76\" (UID: \"f0340c31-098f-4936-bd5e-4ef17c48a9f1\") " pod="openshift-marketplace/redhat-marketplace-rzc76" Oct 08 00:08:50 crc kubenswrapper[4967]: I1008 00:08:50.790595 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f0340c31-098f-4936-bd5e-4ef17c48a9f1-catalog-content\") pod \"redhat-marketplace-rzc76\" (UID: \"f0340c31-098f-4936-bd5e-4ef17c48a9f1\") " pod="openshift-marketplace/redhat-marketplace-rzc76" Oct 08 00:08:50 crc kubenswrapper[4967]: I1008 00:08:50.790646 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fft4x\" (UniqueName: \"kubernetes.io/projected/f0340c31-098f-4936-bd5e-4ef17c48a9f1-kube-api-access-fft4x\") pod \"redhat-marketplace-rzc76\" (UID: \"f0340c31-098f-4936-bd5e-4ef17c48a9f1\") " pod="openshift-marketplace/redhat-marketplace-rzc76" Oct 08 00:08:50 crc kubenswrapper[4967]: I1008 00:08:50.790703 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8d8x5\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:08:50 crc kubenswrapper[4967]: E1008 00:08:50.791082 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 00:08:51.29106693 +0000 UTC m=+129.941601569 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8d8x5" (UID: "1f06a5d1-8ff8-4d73-8939-0fac4252ea9c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:50 crc kubenswrapper[4967]: I1008 00:08:50.891577 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 00:08:50 crc kubenswrapper[4967]: I1008 00:08:50.891851 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f0340c31-098f-4936-bd5e-4ef17c48a9f1-catalog-content\") pod \"redhat-marketplace-rzc76\" (UID: \"f0340c31-098f-4936-bd5e-4ef17c48a9f1\") " pod="openshift-marketplace/redhat-marketplace-rzc76" Oct 08 00:08:50 crc kubenswrapper[4967]: I1008 00:08:50.891931 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fft4x\" (UniqueName: \"kubernetes.io/projected/f0340c31-098f-4936-bd5e-4ef17c48a9f1-kube-api-access-fft4x\") pod \"redhat-marketplace-rzc76\" (UID: \"f0340c31-098f-4936-bd5e-4ef17c48a9f1\") " pod="openshift-marketplace/redhat-marketplace-rzc76" Oct 08 00:08:50 crc kubenswrapper[4967]: I1008 00:08:50.892005 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f0340c31-098f-4936-bd5e-4ef17c48a9f1-utilities\") pod \"redhat-marketplace-rzc76\" (UID: \"f0340c31-098f-4936-bd5e-4ef17c48a9f1\") " pod="openshift-marketplace/redhat-marketplace-rzc76" Oct 08 00:08:50 crc kubenswrapper[4967]: E1008 00:08:50.892567 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 00:08:51.392537887 +0000 UTC m=+130.043072536 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:50 crc kubenswrapper[4967]: I1008 00:08:50.892873 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f0340c31-098f-4936-bd5e-4ef17c48a9f1-utilities\") pod \"redhat-marketplace-rzc76\" (UID: \"f0340c31-098f-4936-bd5e-4ef17c48a9f1\") " pod="openshift-marketplace/redhat-marketplace-rzc76" Oct 08 00:08:50 crc kubenswrapper[4967]: I1008 00:08:50.892954 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f0340c31-098f-4936-bd5e-4ef17c48a9f1-catalog-content\") pod \"redhat-marketplace-rzc76\" (UID: \"f0340c31-098f-4936-bd5e-4ef17c48a9f1\") " pod="openshift-marketplace/redhat-marketplace-rzc76" Oct 08 00:08:50 crc kubenswrapper[4967]: I1008 00:08:50.893882 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 08 00:08:50 crc kubenswrapper[4967]: I1008 00:08:50.894726 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 08 00:08:50 crc kubenswrapper[4967]: I1008 00:08:50.897481 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Oct 08 00:08:50 crc kubenswrapper[4967]: I1008 00:08:50.897703 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Oct 08 00:08:50 crc kubenswrapper[4967]: I1008 00:08:50.907553 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 08 00:08:50 crc kubenswrapper[4967]: I1008 00:08:50.923093 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fft4x\" (UniqueName: \"kubernetes.io/projected/f0340c31-098f-4936-bd5e-4ef17c48a9f1-kube-api-access-fft4x\") pod \"redhat-marketplace-rzc76\" (UID: \"f0340c31-098f-4936-bd5e-4ef17c48a9f1\") " pod="openshift-marketplace/redhat-marketplace-rzc76" Oct 08 00:08:50 crc kubenswrapper[4967]: I1008 00:08:50.971322 4967 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Oct 08 00:08:50 crc kubenswrapper[4967]: I1008 00:08:50.995835 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ea71a082-3a10-42af-b39c-69ec1c0eb242-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"ea71a082-3a10-42af-b39c-69ec1c0eb242\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 08 00:08:50 crc kubenswrapper[4967]: I1008 00:08:50.995913 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ea71a082-3a10-42af-b39c-69ec1c0eb242-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"ea71a082-3a10-42af-b39c-69ec1c0eb242\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 08 00:08:50 crc kubenswrapper[4967]: I1008 00:08:50.995985 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8d8x5\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:08:50 crc kubenswrapper[4967]: E1008 00:08:50.996345 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 00:08:51.496322662 +0000 UTC m=+130.146857311 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8d8x5" (UID: "1f06a5d1-8ff8-4d73-8939-0fac4252ea9c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:51 crc kubenswrapper[4967]: I1008 00:08:51.000495 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rzc76" Oct 08 00:08:51 crc kubenswrapper[4967]: I1008 00:08:51.064752 4967 generic.go:334] "Generic (PLEG): container finished" podID="a0f7251c-fe97-4f8d-8121-82f7010170ef" containerID="62a0ec0ccaa70cfd198c426505e9901fdfc7ea35087b3cd9f7a306a7e954b9e1" exitCode=0 Oct 08 00:08:51 crc kubenswrapper[4967]: I1008 00:08:51.065399 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9sp7f" event={"ID":"a0f7251c-fe97-4f8d-8121-82f7010170ef","Type":"ContainerDied","Data":"62a0ec0ccaa70cfd198c426505e9901fdfc7ea35087b3cd9f7a306a7e954b9e1"} Oct 08 00:08:51 crc kubenswrapper[4967]: I1008 00:08:51.065428 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9sp7f" event={"ID":"a0f7251c-fe97-4f8d-8121-82f7010170ef","Type":"ContainerStarted","Data":"37e06fc47d07f2380a95cc6e276b8ffad84084d333e5067b79ba0544629f4642"} Oct 08 00:08:51 crc kubenswrapper[4967]: I1008 00:08:51.078129 4967 generic.go:334] "Generic (PLEG): container finished" podID="32d2d299-2970-40e9-9116-74a074c80eb1" containerID="0425a89b764948462abab2813a5e0a67578e7f549efe81290c8da7aaa8337561" exitCode=0 Oct 08 00:08:51 crc kubenswrapper[4967]: I1008 00:08:51.078394 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x8d4p" event={"ID":"32d2d299-2970-40e9-9116-74a074c80eb1","Type":"ContainerDied","Data":"0425a89b764948462abab2813a5e0a67578e7f549efe81290c8da7aaa8337561"} Oct 08 00:08:51 crc kubenswrapper[4967]: I1008 00:08:51.093277 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-wcll6"] Oct 08 00:08:51 crc kubenswrapper[4967]: I1008 00:08:51.095148 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wcll6" Oct 08 00:08:51 crc kubenswrapper[4967]: I1008 00:08:51.096661 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 00:08:51 crc kubenswrapper[4967]: I1008 00:08:51.097082 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ea71a082-3a10-42af-b39c-69ec1c0eb242-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"ea71a082-3a10-42af-b39c-69ec1c0eb242\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 08 00:08:51 crc kubenswrapper[4967]: I1008 00:08:51.097161 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ea71a082-3a10-42af-b39c-69ec1c0eb242-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"ea71a082-3a10-42af-b39c-69ec1c0eb242\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 08 00:08:51 crc kubenswrapper[4967]: E1008 00:08:51.097656 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 00:08:51.597638915 +0000 UTC m=+130.248173564 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:51 crc kubenswrapper[4967]: I1008 00:08:51.097712 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ea71a082-3a10-42af-b39c-69ec1c0eb242-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"ea71a082-3a10-42af-b39c-69ec1c0eb242\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 08 00:08:51 crc kubenswrapper[4967]: I1008 00:08:51.113433 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wcll6"] Oct 08 00:08:51 crc kubenswrapper[4967]: I1008 00:08:51.149069 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-t65kf" event={"ID":"aa6ab9dd-70bc-4834-a3be-f934e73c15c5","Type":"ContainerStarted","Data":"bed2d8b5f130a38947f901934a0d9219cc2b938e554b6cd9ab84e798456d35ce"} Oct 08 00:08:51 crc kubenswrapper[4967]: I1008 00:08:51.149755 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ea71a082-3a10-42af-b39c-69ec1c0eb242-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"ea71a082-3a10-42af-b39c-69ec1c0eb242\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 08 00:08:51 crc kubenswrapper[4967]: I1008 00:08:51.167501 4967 generic.go:334] "Generic (PLEG): container finished" podID="8912b04d-fd70-497c-aab6-7490fc6ddf40" containerID="5ae6a42d62117bc7aeca9ebb2c6aa58dc5c18e8f8bbabcbb287f458de8c03e89" exitCode=0 Oct 08 00:08:51 crc kubenswrapper[4967]: I1008 00:08:51.167821 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4hhcd" event={"ID":"8912b04d-fd70-497c-aab6-7490fc6ddf40","Type":"ContainerDied","Data":"5ae6a42d62117bc7aeca9ebb2c6aa58dc5c18e8f8bbabcbb287f458de8c03e89"} Oct 08 00:08:51 crc kubenswrapper[4967]: I1008 00:08:51.167862 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4hhcd" event={"ID":"8912b04d-fd70-497c-aab6-7490fc6ddf40","Type":"ContainerStarted","Data":"719b9ed2d135e5117f32eefac6f2b3426f69bddf531129a000cfc6e41aeb0848"} Oct 08 00:08:51 crc kubenswrapper[4967]: I1008 00:08:51.198275 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c827bbc-b90c-4168-918d-c2658da9b6ac-catalog-content\") pod \"redhat-marketplace-wcll6\" (UID: \"9c827bbc-b90c-4168-918d-c2658da9b6ac\") " pod="openshift-marketplace/redhat-marketplace-wcll6" Oct 08 00:08:51 crc kubenswrapper[4967]: I1008 00:08:51.198681 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8d8x5\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:08:51 crc kubenswrapper[4967]: I1008 00:08:51.198713 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c827bbc-b90c-4168-918d-c2658da9b6ac-utilities\") pod \"redhat-marketplace-wcll6\" (UID: \"9c827bbc-b90c-4168-918d-c2658da9b6ac\") " pod="openshift-marketplace/redhat-marketplace-wcll6" Oct 08 00:08:51 crc kubenswrapper[4967]: I1008 00:08:51.198763 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zq8z8\" (UniqueName: \"kubernetes.io/projected/9c827bbc-b90c-4168-918d-c2658da9b6ac-kube-api-access-zq8z8\") pod \"redhat-marketplace-wcll6\" (UID: \"9c827bbc-b90c-4168-918d-c2658da9b6ac\") " pod="openshift-marketplace/redhat-marketplace-wcll6" Oct 08 00:08:51 crc kubenswrapper[4967]: E1008 00:08:51.199783 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 00:08:51.699770188 +0000 UTC m=+130.350304827 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8d8x5" (UID: "1f06a5d1-8ff8-4d73-8939-0fac4252ea9c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:51 crc kubenswrapper[4967]: I1008 00:08:51.230995 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 08 00:08:51 crc kubenswrapper[4967]: I1008 00:08:51.297160 4967 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-10-08T00:08:50.971353133Z","Handler":null,"Name":""} Oct 08 00:08:51 crc kubenswrapper[4967]: I1008 00:08:51.300949 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 00:08:51 crc kubenswrapper[4967]: E1008 00:08:51.301144 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 00:08:51.801101861 +0000 UTC m=+130.451636510 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:51 crc kubenswrapper[4967]: I1008 00:08:51.301445 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8d8x5\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:08:51 crc kubenswrapper[4967]: I1008 00:08:51.301503 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c827bbc-b90c-4168-918d-c2658da9b6ac-utilities\") pod \"redhat-marketplace-wcll6\" (UID: \"9c827bbc-b90c-4168-918d-c2658da9b6ac\") " pod="openshift-marketplace/redhat-marketplace-wcll6" Oct 08 00:08:51 crc kubenswrapper[4967]: I1008 00:08:51.301625 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zq8z8\" (UniqueName: \"kubernetes.io/projected/9c827bbc-b90c-4168-918d-c2658da9b6ac-kube-api-access-zq8z8\") pod \"redhat-marketplace-wcll6\" (UID: \"9c827bbc-b90c-4168-918d-c2658da9b6ac\") " pod="openshift-marketplace/redhat-marketplace-wcll6" Oct 08 00:08:51 crc kubenswrapper[4967]: I1008 00:08:51.301842 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c827bbc-b90c-4168-918d-c2658da9b6ac-catalog-content\") pod \"redhat-marketplace-wcll6\" (UID: \"9c827bbc-b90c-4168-918d-c2658da9b6ac\") " pod="openshift-marketplace/redhat-marketplace-wcll6" Oct 08 00:08:51 crc kubenswrapper[4967]: I1008 00:08:51.302368 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c827bbc-b90c-4168-918d-c2658da9b6ac-catalog-content\") pod \"redhat-marketplace-wcll6\" (UID: \"9c827bbc-b90c-4168-918d-c2658da9b6ac\") " pod="openshift-marketplace/redhat-marketplace-wcll6" Oct 08 00:08:51 crc kubenswrapper[4967]: E1008 00:08:51.302711 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 00:08:51.802702841 +0000 UTC m=+130.453237490 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8d8x5" (UID: "1f06a5d1-8ff8-4d73-8939-0fac4252ea9c") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 00:08:51 crc kubenswrapper[4967]: I1008 00:08:51.304419 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c827bbc-b90c-4168-918d-c2658da9b6ac-utilities\") pod \"redhat-marketplace-wcll6\" (UID: \"9c827bbc-b90c-4168-918d-c2658da9b6ac\") " pod="openshift-marketplace/redhat-marketplace-wcll6" Oct 08 00:08:51 crc kubenswrapper[4967]: I1008 00:08:51.314333 4967 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Oct 08 00:08:51 crc kubenswrapper[4967]: I1008 00:08:51.314374 4967 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Oct 08 00:08:51 crc kubenswrapper[4967]: I1008 00:08:51.351040 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zq8z8\" (UniqueName: \"kubernetes.io/projected/9c827bbc-b90c-4168-918d-c2658da9b6ac-kube-api-access-zq8z8\") pod \"redhat-marketplace-wcll6\" (UID: \"9c827bbc-b90c-4168-918d-c2658da9b6ac\") " pod="openshift-marketplace/redhat-marketplace-wcll6" Oct 08 00:08:51 crc kubenswrapper[4967]: I1008 00:08:51.399329 4967 patch_prober.go:28] interesting pod/router-default-5444994796-p9kxf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 00:08:51 crc kubenswrapper[4967]: [-]has-synced failed: reason withheld Oct 08 00:08:51 crc kubenswrapper[4967]: [+]process-running ok Oct 08 00:08:51 crc kubenswrapper[4967]: healthz check failed Oct 08 00:08:51 crc kubenswrapper[4967]: I1008 00:08:51.399396 4967 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-p9kxf" podUID="ee7dadb3-3ab7-4790-9b67-7f7a5bf7604f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 00:08:51 crc kubenswrapper[4967]: I1008 00:08:51.402300 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rzc76"] Oct 08 00:08:51 crc kubenswrapper[4967]: I1008 00:08:51.407982 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 00:08:51 crc kubenswrapper[4967]: I1008 00:08:51.418990 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 08 00:08:51 crc kubenswrapper[4967]: I1008 00:08:51.419473 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wcll6" Oct 08 00:08:51 crc kubenswrapper[4967]: I1008 00:08:51.510643 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8d8x5\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:08:51 crc kubenswrapper[4967]: I1008 00:08:51.514775 4967 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 08 00:08:51 crc kubenswrapper[4967]: I1008 00:08:51.514809 4967 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8d8x5\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:08:51 crc kubenswrapper[4967]: I1008 00:08:51.608811 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8d8x5\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:08:51 crc kubenswrapper[4967]: I1008 00:08:51.615925 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 08 00:08:51 crc kubenswrapper[4967]: I1008 00:08:51.682712 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-v7scw"] Oct 08 00:08:51 crc kubenswrapper[4967]: I1008 00:08:51.689527 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-v7scw" Oct 08 00:08:51 crc kubenswrapper[4967]: I1008 00:08:51.694918 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 08 00:08:51 crc kubenswrapper[4967]: I1008 00:08:51.700798 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-v7scw"] Oct 08 00:08:51 crc kubenswrapper[4967]: I1008 00:08:51.757861 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wcll6"] Oct 08 00:08:51 crc kubenswrapper[4967]: W1008 00:08:51.781228 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9c827bbc_b90c_4168_918d_c2658da9b6ac.slice/crio-8a6cbc39593c3932a978b35eb4d2adeeca4580b1fe025b83427766b4d5c28ae1 WatchSource:0}: Error finding container 8a6cbc39593c3932a978b35eb4d2adeeca4580b1fe025b83427766b4d5c28ae1: Status 404 returned error can't find the container with id 8a6cbc39593c3932a978b35eb4d2adeeca4580b1fe025b83427766b4d5c28ae1 Oct 08 00:08:51 crc kubenswrapper[4967]: I1008 00:08:51.816409 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5qq8z\" (UniqueName: \"kubernetes.io/projected/f3001d43-c530-4c5e-a6e6-27346682ba5c-kube-api-access-5qq8z\") pod \"redhat-operators-v7scw\" (UID: \"f3001d43-c530-4c5e-a6e6-27346682ba5c\") " pod="openshift-marketplace/redhat-operators-v7scw" Oct 08 00:08:51 crc kubenswrapper[4967]: I1008 00:08:51.816460 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3001d43-c530-4c5e-a6e6-27346682ba5c-utilities\") pod \"redhat-operators-v7scw\" (UID: \"f3001d43-c530-4c5e-a6e6-27346682ba5c\") " pod="openshift-marketplace/redhat-operators-v7scw" Oct 08 00:08:51 crc kubenswrapper[4967]: I1008 00:08:51.816482 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3001d43-c530-4c5e-a6e6-27346682ba5c-catalog-content\") pod \"redhat-operators-v7scw\" (UID: \"f3001d43-c530-4c5e-a6e6-27346682ba5c\") " pod="openshift-marketplace/redhat-operators-v7scw" Oct 08 00:08:51 crc kubenswrapper[4967]: I1008 00:08:51.903810 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Oct 08 00:08:51 crc kubenswrapper[4967]: I1008 00:08:51.913418 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:08:51 crc kubenswrapper[4967]: I1008 00:08:51.918331 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5qq8z\" (UniqueName: \"kubernetes.io/projected/f3001d43-c530-4c5e-a6e6-27346682ba5c-kube-api-access-5qq8z\") pod \"redhat-operators-v7scw\" (UID: \"f3001d43-c530-4c5e-a6e6-27346682ba5c\") " pod="openshift-marketplace/redhat-operators-v7scw" Oct 08 00:08:51 crc kubenswrapper[4967]: I1008 00:08:51.918384 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3001d43-c530-4c5e-a6e6-27346682ba5c-utilities\") pod \"redhat-operators-v7scw\" (UID: \"f3001d43-c530-4c5e-a6e6-27346682ba5c\") " pod="openshift-marketplace/redhat-operators-v7scw" Oct 08 00:08:51 crc kubenswrapper[4967]: I1008 00:08:51.918429 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3001d43-c530-4c5e-a6e6-27346682ba5c-catalog-content\") pod \"redhat-operators-v7scw\" (UID: \"f3001d43-c530-4c5e-a6e6-27346682ba5c\") " pod="openshift-marketplace/redhat-operators-v7scw" Oct 08 00:08:51 crc kubenswrapper[4967]: I1008 00:08:51.919464 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3001d43-c530-4c5e-a6e6-27346682ba5c-catalog-content\") pod \"redhat-operators-v7scw\" (UID: \"f3001d43-c530-4c5e-a6e6-27346682ba5c\") " pod="openshift-marketplace/redhat-operators-v7scw" Oct 08 00:08:51 crc kubenswrapper[4967]: I1008 00:08:51.919603 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3001d43-c530-4c5e-a6e6-27346682ba5c-utilities\") pod \"redhat-operators-v7scw\" (UID: \"f3001d43-c530-4c5e-a6e6-27346682ba5c\") " pod="openshift-marketplace/redhat-operators-v7scw" Oct 08 00:08:51 crc kubenswrapper[4967]: I1008 00:08:51.942168 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5qq8z\" (UniqueName: \"kubernetes.io/projected/f3001d43-c530-4c5e-a6e6-27346682ba5c-kube-api-access-5qq8z\") pod \"redhat-operators-v7scw\" (UID: \"f3001d43-c530-4c5e-a6e6-27346682ba5c\") " pod="openshift-marketplace/redhat-operators-v7scw" Oct 08 00:08:51 crc kubenswrapper[4967]: I1008 00:08:51.972787 4967 patch_prober.go:28] interesting pod/downloads-7954f5f757-tdqfv container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.18:8080/\": dial tcp 10.217.0.18:8080: connect: connection refused" start-of-body= Oct 08 00:08:51 crc kubenswrapper[4967]: I1008 00:08:51.973172 4967 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-tdqfv" podUID="fa677745-a987-479c-b60a-2220f1bd58d1" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.18:8080/\": dial tcp 10.217.0.18:8080: connect: connection refused" Oct 08 00:08:51 crc kubenswrapper[4967]: I1008 00:08:51.979212 4967 patch_prober.go:28] interesting pod/downloads-7954f5f757-tdqfv container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.18:8080/\": dial tcp 10.217.0.18:8080: connect: connection refused" start-of-body= Oct 08 00:08:51 crc kubenswrapper[4967]: I1008 00:08:51.979267 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-tdqfv" podUID="fa677745-a987-479c-b60a-2220f1bd58d1" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.18:8080/\": dial tcp 10.217.0.18:8080: connect: connection refused" Oct 08 00:08:52 crc kubenswrapper[4967]: I1008 00:08:52.021877 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-v7scw" Oct 08 00:08:52 crc kubenswrapper[4967]: I1008 00:08:52.085425 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-f24p2"] Oct 08 00:08:52 crc kubenswrapper[4967]: I1008 00:08:52.086764 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f24p2" Oct 08 00:08:52 crc kubenswrapper[4967]: I1008 00:08:52.099805 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-f24p2"] Oct 08 00:08:52 crc kubenswrapper[4967]: I1008 00:08:52.129544 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Oct 08 00:08:52 crc kubenswrapper[4967]: I1008 00:08:52.138997 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nf2kn\" (UniqueName: \"kubernetes.io/projected/92a4f64f-e1cb-42a1-85b4-d3fc38e949a2-kube-api-access-nf2kn\") pod \"redhat-operators-f24p2\" (UID: \"92a4f64f-e1cb-42a1-85b4-d3fc38e949a2\") " pod="openshift-marketplace/redhat-operators-f24p2" Oct 08 00:08:52 crc kubenswrapper[4967]: I1008 00:08:52.139122 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92a4f64f-e1cb-42a1-85b4-d3fc38e949a2-utilities\") pod \"redhat-operators-f24p2\" (UID: \"92a4f64f-e1cb-42a1-85b4-d3fc38e949a2\") " pod="openshift-marketplace/redhat-operators-f24p2" Oct 08 00:08:52 crc kubenswrapper[4967]: I1008 00:08:52.139195 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92a4f64f-e1cb-42a1-85b4-d3fc38e949a2-catalog-content\") pod \"redhat-operators-f24p2\" (UID: \"92a4f64f-e1cb-42a1-85b4-d3fc38e949a2\") " pod="openshift-marketplace/redhat-operators-f24p2" Oct 08 00:08:52 crc kubenswrapper[4967]: I1008 00:08:52.186430 4967 generic.go:334] "Generic (PLEG): container finished" podID="f0340c31-098f-4936-bd5e-4ef17c48a9f1" containerID="6e8c3a1ca8260e2aff066ffa90cac9d502bed952d2d7901effc492dabd1c3702" exitCode=0 Oct 08 00:08:52 crc kubenswrapper[4967]: I1008 00:08:52.186534 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rzc76" event={"ID":"f0340c31-098f-4936-bd5e-4ef17c48a9f1","Type":"ContainerDied","Data":"6e8c3a1ca8260e2aff066ffa90cac9d502bed952d2d7901effc492dabd1c3702"} Oct 08 00:08:52 crc kubenswrapper[4967]: I1008 00:08:52.186576 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rzc76" event={"ID":"f0340c31-098f-4936-bd5e-4ef17c48a9f1","Type":"ContainerStarted","Data":"cf7ef634599a2521aeb21dcf825d637113880f802e35fa8ebeeed633ffb7d906"} Oct 08 00:08:52 crc kubenswrapper[4967]: I1008 00:08:52.201113 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-t65kf" event={"ID":"aa6ab9dd-70bc-4834-a3be-f934e73c15c5","Type":"ContainerStarted","Data":"6f9483b81cccc7b33598599e171995eae11901dd1120ddb50b31d63a7095d0e6"} Oct 08 00:08:52 crc kubenswrapper[4967]: I1008 00:08:52.201231 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-t65kf" event={"ID":"aa6ab9dd-70bc-4834-a3be-f934e73c15c5","Type":"ContainerStarted","Data":"73a4b889934d600c36fd4dad153de0742ca7a0a69159f40c147f20948d868f08"} Oct 08 00:08:52 crc kubenswrapper[4967]: I1008 00:08:52.206535 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"ea71a082-3a10-42af-b39c-69ec1c0eb242","Type":"ContainerStarted","Data":"f3104984097ec392438cf7eefed160ac3ba7485eb92e23a2ce2cd1a1cca34c32"} Oct 08 00:08:52 crc kubenswrapper[4967]: I1008 00:08:52.234085 4967 generic.go:334] "Generic (PLEG): container finished" podID="9c827bbc-b90c-4168-918d-c2658da9b6ac" containerID="ed2206f1704b83fde92a9711faa9784858d66a94534908ef494a9a4864ba0a93" exitCode=0 Oct 08 00:08:52 crc kubenswrapper[4967]: I1008 00:08:52.234123 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wcll6" event={"ID":"9c827bbc-b90c-4168-918d-c2658da9b6ac","Type":"ContainerDied","Data":"ed2206f1704b83fde92a9711faa9784858d66a94534908ef494a9a4864ba0a93"} Oct 08 00:08:52 crc kubenswrapper[4967]: I1008 00:08:52.234147 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wcll6" event={"ID":"9c827bbc-b90c-4168-918d-c2658da9b6ac","Type":"ContainerStarted","Data":"8a6cbc39593c3932a978b35eb4d2adeeca4580b1fe025b83427766b4d5c28ae1"} Oct 08 00:08:52 crc kubenswrapper[4967]: I1008 00:08:52.239972 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92a4f64f-e1cb-42a1-85b4-d3fc38e949a2-utilities\") pod \"redhat-operators-f24p2\" (UID: \"92a4f64f-e1cb-42a1-85b4-d3fc38e949a2\") " pod="openshift-marketplace/redhat-operators-f24p2" Oct 08 00:08:52 crc kubenswrapper[4967]: I1008 00:08:52.240032 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92a4f64f-e1cb-42a1-85b4-d3fc38e949a2-catalog-content\") pod \"redhat-operators-f24p2\" (UID: \"92a4f64f-e1cb-42a1-85b4-d3fc38e949a2\") " pod="openshift-marketplace/redhat-operators-f24p2" Oct 08 00:08:52 crc kubenswrapper[4967]: I1008 00:08:52.240103 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nf2kn\" (UniqueName: \"kubernetes.io/projected/92a4f64f-e1cb-42a1-85b4-d3fc38e949a2-kube-api-access-nf2kn\") pod \"redhat-operators-f24p2\" (UID: \"92a4f64f-e1cb-42a1-85b4-d3fc38e949a2\") " pod="openshift-marketplace/redhat-operators-f24p2" Oct 08 00:08:52 crc kubenswrapper[4967]: I1008 00:08:52.240859 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92a4f64f-e1cb-42a1-85b4-d3fc38e949a2-utilities\") pod \"redhat-operators-f24p2\" (UID: \"92a4f64f-e1cb-42a1-85b4-d3fc38e949a2\") " pod="openshift-marketplace/redhat-operators-f24p2" Oct 08 00:08:52 crc kubenswrapper[4967]: I1008 00:08:52.241067 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92a4f64f-e1cb-42a1-85b4-d3fc38e949a2-catalog-content\") pod \"redhat-operators-f24p2\" (UID: \"92a4f64f-e1cb-42a1-85b4-d3fc38e949a2\") " pod="openshift-marketplace/redhat-operators-f24p2" Oct 08 00:08:52 crc kubenswrapper[4967]: I1008 00:08:52.262594 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-t65kf" podStartSLOduration=13.262568829 podStartE2EDuration="13.262568829s" podCreationTimestamp="2025-10-08 00:08:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:08:52.258175599 +0000 UTC m=+130.908710248" watchObservedRunningTime="2025-10-08 00:08:52.262568829 +0000 UTC m=+130.913103478" Oct 08 00:08:52 crc kubenswrapper[4967]: I1008 00:08:52.268847 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nf2kn\" (UniqueName: \"kubernetes.io/projected/92a4f64f-e1cb-42a1-85b4-d3fc38e949a2-kube-api-access-nf2kn\") pod \"redhat-operators-f24p2\" (UID: \"92a4f64f-e1cb-42a1-85b4-d3fc38e949a2\") " pod="openshift-marketplace/redhat-operators-f24p2" Oct 08 00:08:52 crc kubenswrapper[4967]: I1008 00:08:52.394495 4967 patch_prober.go:28] interesting pod/router-default-5444994796-p9kxf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 00:08:52 crc kubenswrapper[4967]: [-]has-synced failed: reason withheld Oct 08 00:08:52 crc kubenswrapper[4967]: [+]process-running ok Oct 08 00:08:52 crc kubenswrapper[4967]: healthz check failed Oct 08 00:08:52 crc kubenswrapper[4967]: I1008 00:08:52.397598 4967 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-p9kxf" podUID="ee7dadb3-3ab7-4790-9b67-7f7a5bf7604f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 00:08:52 crc kubenswrapper[4967]: I1008 00:08:52.446080 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-ck4pd" Oct 08 00:08:52 crc kubenswrapper[4967]: I1008 00:08:52.452929 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-ck4pd" Oct 08 00:08:52 crc kubenswrapper[4967]: I1008 00:08:52.491499 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f24p2" Oct 08 00:08:52 crc kubenswrapper[4967]: I1008 00:08:52.577565 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-8d8x5"] Oct 08 00:08:52 crc kubenswrapper[4967]: I1008 00:08:52.644481 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-v7scw"] Oct 08 00:08:52 crc kubenswrapper[4967]: W1008 00:08:52.654360 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf3001d43_c530_4c5e_a6e6_27346682ba5c.slice/crio-79f4a8d4284a11c7df546ef08ef87f4a94b04fe0b759ceb5097711008c24bcfa WatchSource:0}: Error finding container 79f4a8d4284a11c7df546ef08ef87f4a94b04fe0b759ceb5097711008c24bcfa: Status 404 returned error can't find the container with id 79f4a8d4284a11c7df546ef08ef87f4a94b04fe0b759ceb5097711008c24bcfa Oct 08 00:08:52 crc kubenswrapper[4967]: I1008 00:08:52.827986 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-f24p2"] Oct 08 00:08:52 crc kubenswrapper[4967]: W1008 00:08:52.844603 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod92a4f64f_e1cb_42a1_85b4_d3fc38e949a2.slice/crio-5765d732093b5e0be047c8db149b7dbb5008bcb5ed58e81a6284b667cbb7121b WatchSource:0}: Error finding container 5765d732093b5e0be047c8db149b7dbb5008bcb5ed58e81a6284b667cbb7121b: Status 404 returned error can't find the container with id 5765d732093b5e0be047c8db149b7dbb5008bcb5ed58e81a6284b667cbb7121b Oct 08 00:08:53 crc kubenswrapper[4967]: E1008 00:08:53.092790 4967 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf3001d43_c530_4c5e_a6e6_27346682ba5c.slice/crio-465c13717c0d1dc7774b70e329714c01d9a30e91684ef10a3fa5ffa232423eab.scope\": RecentStats: unable to find data in memory cache]" Oct 08 00:08:53 crc kubenswrapper[4967]: I1008 00:08:53.245344 4967 generic.go:334] "Generic (PLEG): container finished" podID="bb260fdf-16d6-4771-b7ae-38a669ff2393" containerID="7782cfe4b15523fd5b3b4d5ad87a4a2bf3d12cfca08fc3be11370557a2830069" exitCode=0 Oct 08 00:08:53 crc kubenswrapper[4967]: I1008 00:08:53.245480 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331360-t4jlf" event={"ID":"bb260fdf-16d6-4771-b7ae-38a669ff2393","Type":"ContainerDied","Data":"7782cfe4b15523fd5b3b4d5ad87a4a2bf3d12cfca08fc3be11370557a2830069"} Oct 08 00:08:53 crc kubenswrapper[4967]: I1008 00:08:53.253187 4967 generic.go:334] "Generic (PLEG): container finished" podID="ea71a082-3a10-42af-b39c-69ec1c0eb242" containerID="f5706577c5efe1f1c0e94b248bf4beaf0f86e3183e0cb3f2a985e74d7029f606" exitCode=0 Oct 08 00:08:53 crc kubenswrapper[4967]: I1008 00:08:53.253289 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"ea71a082-3a10-42af-b39c-69ec1c0eb242","Type":"ContainerDied","Data":"f5706577c5efe1f1c0e94b248bf4beaf0f86e3183e0cb3f2a985e74d7029f606"} Oct 08 00:08:53 crc kubenswrapper[4967]: I1008 00:08:53.263817 4967 generic.go:334] "Generic (PLEG): container finished" podID="92a4f64f-e1cb-42a1-85b4-d3fc38e949a2" containerID="ec33944e8ccb6004145beaaa46539b7be44284a9a6c3afd0c283efdc6ab8c569" exitCode=0 Oct 08 00:08:53 crc kubenswrapper[4967]: I1008 00:08:53.263936 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f24p2" event={"ID":"92a4f64f-e1cb-42a1-85b4-d3fc38e949a2","Type":"ContainerDied","Data":"ec33944e8ccb6004145beaaa46539b7be44284a9a6c3afd0c283efdc6ab8c569"} Oct 08 00:08:53 crc kubenswrapper[4967]: I1008 00:08:53.263964 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f24p2" event={"ID":"92a4f64f-e1cb-42a1-85b4-d3fc38e949a2","Type":"ContainerStarted","Data":"5765d732093b5e0be047c8db149b7dbb5008bcb5ed58e81a6284b667cbb7121b"} Oct 08 00:08:53 crc kubenswrapper[4967]: I1008 00:08:53.266460 4967 generic.go:334] "Generic (PLEG): container finished" podID="f3001d43-c530-4c5e-a6e6-27346682ba5c" containerID="465c13717c0d1dc7774b70e329714c01d9a30e91684ef10a3fa5ffa232423eab" exitCode=0 Oct 08 00:08:53 crc kubenswrapper[4967]: I1008 00:08:53.266546 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v7scw" event={"ID":"f3001d43-c530-4c5e-a6e6-27346682ba5c","Type":"ContainerDied","Data":"465c13717c0d1dc7774b70e329714c01d9a30e91684ef10a3fa5ffa232423eab"} Oct 08 00:08:53 crc kubenswrapper[4967]: I1008 00:08:53.266573 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v7scw" event={"ID":"f3001d43-c530-4c5e-a6e6-27346682ba5c","Type":"ContainerStarted","Data":"79f4a8d4284a11c7df546ef08ef87f4a94b04fe0b759ceb5097711008c24bcfa"} Oct 08 00:08:53 crc kubenswrapper[4967]: I1008 00:08:53.272278 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" event={"ID":"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c","Type":"ContainerStarted","Data":"b8d42b81a779b7bfaa854fb91cbc2f3948dc9bc6aac2b6c0fd684e856fc1871e"} Oct 08 00:08:53 crc kubenswrapper[4967]: I1008 00:08:53.272421 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" event={"ID":"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c","Type":"ContainerStarted","Data":"2df3f6aadad5e3c63cc4ae1da8849d34672ca81be52b38a43f9cd0034ab5018c"} Oct 08 00:08:53 crc kubenswrapper[4967]: I1008 00:08:53.272973 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:08:53 crc kubenswrapper[4967]: I1008 00:08:53.351039 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" podStartSLOduration=109.351017135 podStartE2EDuration="1m49.351017135s" podCreationTimestamp="2025-10-08 00:07:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:08:53.347840456 +0000 UTC m=+131.998375115" watchObservedRunningTime="2025-10-08 00:08:53.351017135 +0000 UTC m=+132.001551784" Oct 08 00:08:53 crc kubenswrapper[4967]: I1008 00:08:53.394039 4967 patch_prober.go:28] interesting pod/router-default-5444994796-p9kxf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 00:08:53 crc kubenswrapper[4967]: [-]has-synced failed: reason withheld Oct 08 00:08:53 crc kubenswrapper[4967]: [+]process-running ok Oct 08 00:08:53 crc kubenswrapper[4967]: healthz check failed Oct 08 00:08:53 crc kubenswrapper[4967]: I1008 00:08:53.394102 4967 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-p9kxf" podUID="ee7dadb3-3ab7-4790-9b67-7f7a5bf7604f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 00:08:53 crc kubenswrapper[4967]: I1008 00:08:53.444907 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 08 00:08:53 crc kubenswrapper[4967]: I1008 00:08:53.449133 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 08 00:08:53 crc kubenswrapper[4967]: I1008 00:08:53.452434 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Oct 08 00:08:53 crc kubenswrapper[4967]: I1008 00:08:53.453027 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Oct 08 00:08:53 crc kubenswrapper[4967]: I1008 00:08:53.467154 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 08 00:08:53 crc kubenswrapper[4967]: I1008 00:08:53.472989 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/85a3da15-9895-4e7f-81f4-6be5e34061de-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"85a3da15-9895-4e7f-81f4-6be5e34061de\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 08 00:08:53 crc kubenswrapper[4967]: I1008 00:08:53.473044 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/85a3da15-9895-4e7f-81f4-6be5e34061de-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"85a3da15-9895-4e7f-81f4-6be5e34061de\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 08 00:08:53 crc kubenswrapper[4967]: I1008 00:08:53.574446 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/85a3da15-9895-4e7f-81f4-6be5e34061de-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"85a3da15-9895-4e7f-81f4-6be5e34061de\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 08 00:08:53 crc kubenswrapper[4967]: I1008 00:08:53.574515 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/85a3da15-9895-4e7f-81f4-6be5e34061de-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"85a3da15-9895-4e7f-81f4-6be5e34061de\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 08 00:08:53 crc kubenswrapper[4967]: I1008 00:08:53.574807 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/85a3da15-9895-4e7f-81f4-6be5e34061de-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"85a3da15-9895-4e7f-81f4-6be5e34061de\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 08 00:08:53 crc kubenswrapper[4967]: I1008 00:08:53.604643 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/85a3da15-9895-4e7f-81f4-6be5e34061de-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"85a3da15-9895-4e7f-81f4-6be5e34061de\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 08 00:08:53 crc kubenswrapper[4967]: I1008 00:08:53.782882 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 08 00:08:54 crc kubenswrapper[4967]: I1008 00:08:54.400490 4967 patch_prober.go:28] interesting pod/router-default-5444994796-p9kxf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 00:08:54 crc kubenswrapper[4967]: [-]has-synced failed: reason withheld Oct 08 00:08:54 crc kubenswrapper[4967]: [+]process-running ok Oct 08 00:08:54 crc kubenswrapper[4967]: healthz check failed Oct 08 00:08:54 crc kubenswrapper[4967]: I1008 00:08:54.400947 4967 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-p9kxf" podUID="ee7dadb3-3ab7-4790-9b67-7f7a5bf7604f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 00:08:54 crc kubenswrapper[4967]: I1008 00:08:54.436789 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 08 00:08:54 crc kubenswrapper[4967]: I1008 00:08:54.698088 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-vxdxg" Oct 08 00:08:54 crc kubenswrapper[4967]: I1008 00:08:54.698822 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-vxdxg" Oct 08 00:08:54 crc kubenswrapper[4967]: I1008 00:08:54.701351 4967 patch_prober.go:28] interesting pod/console-f9d7485db-vxdxg container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.17:8443/health\": dial tcp 10.217.0.17:8443: connect: connection refused" start-of-body= Oct 08 00:08:54 crc kubenswrapper[4967]: I1008 00:08:54.701392 4967 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-vxdxg" podUID="8e5ce93e-573e-456c-a08c-4737fb3c7a66" containerName="console" probeResult="failure" output="Get \"https://10.217.0.17:8443/health\": dial tcp 10.217.0.17:8443: connect: connection refused" Oct 08 00:08:54 crc kubenswrapper[4967]: I1008 00:08:54.723726 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331360-t4jlf" Oct 08 00:08:54 crc kubenswrapper[4967]: I1008 00:08:54.723740 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 08 00:08:54 crc kubenswrapper[4967]: I1008 00:08:54.850078 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-npdfq" Oct 08 00:08:54 crc kubenswrapper[4967]: I1008 00:08:54.911614 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ea71a082-3a10-42af-b39c-69ec1c0eb242-kubelet-dir\") pod \"ea71a082-3a10-42af-b39c-69ec1c0eb242\" (UID: \"ea71a082-3a10-42af-b39c-69ec1c0eb242\") " Oct 08 00:08:54 crc kubenswrapper[4967]: I1008 00:08:54.911733 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bb260fdf-16d6-4771-b7ae-38a669ff2393-secret-volume\") pod \"bb260fdf-16d6-4771-b7ae-38a669ff2393\" (UID: \"bb260fdf-16d6-4771-b7ae-38a669ff2393\") " Oct 08 00:08:54 crc kubenswrapper[4967]: I1008 00:08:54.911786 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-48bwp\" (UniqueName: \"kubernetes.io/projected/bb260fdf-16d6-4771-b7ae-38a669ff2393-kube-api-access-48bwp\") pod \"bb260fdf-16d6-4771-b7ae-38a669ff2393\" (UID: \"bb260fdf-16d6-4771-b7ae-38a669ff2393\") " Oct 08 00:08:54 crc kubenswrapper[4967]: I1008 00:08:54.911831 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bb260fdf-16d6-4771-b7ae-38a669ff2393-config-volume\") pod \"bb260fdf-16d6-4771-b7ae-38a669ff2393\" (UID: \"bb260fdf-16d6-4771-b7ae-38a669ff2393\") " Oct 08 00:08:54 crc kubenswrapper[4967]: I1008 00:08:54.911834 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ea71a082-3a10-42af-b39c-69ec1c0eb242-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "ea71a082-3a10-42af-b39c-69ec1c0eb242" (UID: "ea71a082-3a10-42af-b39c-69ec1c0eb242"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 00:08:54 crc kubenswrapper[4967]: I1008 00:08:54.911849 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ea71a082-3a10-42af-b39c-69ec1c0eb242-kube-api-access\") pod \"ea71a082-3a10-42af-b39c-69ec1c0eb242\" (UID: \"ea71a082-3a10-42af-b39c-69ec1c0eb242\") " Oct 08 00:08:54 crc kubenswrapper[4967]: I1008 00:08:54.912408 4967 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ea71a082-3a10-42af-b39c-69ec1c0eb242-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 08 00:08:54 crc kubenswrapper[4967]: I1008 00:08:54.914293 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb260fdf-16d6-4771-b7ae-38a669ff2393-config-volume" (OuterVolumeSpecName: "config-volume") pod "bb260fdf-16d6-4771-b7ae-38a669ff2393" (UID: "bb260fdf-16d6-4771-b7ae-38a669ff2393"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:08:54 crc kubenswrapper[4967]: I1008 00:08:54.917972 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb260fdf-16d6-4771-b7ae-38a669ff2393-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "bb260fdf-16d6-4771-b7ae-38a669ff2393" (UID: "bb260fdf-16d6-4771-b7ae-38a669ff2393"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:08:54 crc kubenswrapper[4967]: I1008 00:08:54.919786 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea71a082-3a10-42af-b39c-69ec1c0eb242-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "ea71a082-3a10-42af-b39c-69ec1c0eb242" (UID: "ea71a082-3a10-42af-b39c-69ec1c0eb242"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:08:54 crc kubenswrapper[4967]: I1008 00:08:54.925006 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb260fdf-16d6-4771-b7ae-38a669ff2393-kube-api-access-48bwp" (OuterVolumeSpecName: "kube-api-access-48bwp") pod "bb260fdf-16d6-4771-b7ae-38a669ff2393" (UID: "bb260fdf-16d6-4771-b7ae-38a669ff2393"). InnerVolumeSpecName "kube-api-access-48bwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:08:55 crc kubenswrapper[4967]: I1008 00:08:55.017923 4967 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bb260fdf-16d6-4771-b7ae-38a669ff2393-config-volume\") on node \"crc\" DevicePath \"\"" Oct 08 00:08:55 crc kubenswrapper[4967]: I1008 00:08:55.017958 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ea71a082-3a10-42af-b39c-69ec1c0eb242-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 08 00:08:55 crc kubenswrapper[4967]: I1008 00:08:55.017970 4967 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bb260fdf-16d6-4771-b7ae-38a669ff2393-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 08 00:08:55 crc kubenswrapper[4967]: I1008 00:08:55.017980 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-48bwp\" (UniqueName: \"kubernetes.io/projected/bb260fdf-16d6-4771-b7ae-38a669ff2393-kube-api-access-48bwp\") on node \"crc\" DevicePath \"\"" Oct 08 00:08:55 crc kubenswrapper[4967]: I1008 00:08:55.294204 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331360-t4jlf" event={"ID":"bb260fdf-16d6-4771-b7ae-38a669ff2393","Type":"ContainerDied","Data":"e72b6d516bfd9e6181cbf6d1914738858269573d66ed955b3111e061c3078155"} Oct 08 00:08:55 crc kubenswrapper[4967]: I1008 00:08:55.294485 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e72b6d516bfd9e6181cbf6d1914738858269573d66ed955b3111e061c3078155" Oct 08 00:08:55 crc kubenswrapper[4967]: I1008 00:08:55.294546 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331360-t4jlf" Oct 08 00:08:55 crc kubenswrapper[4967]: I1008 00:08:55.311114 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"ea71a082-3a10-42af-b39c-69ec1c0eb242","Type":"ContainerDied","Data":"f3104984097ec392438cf7eefed160ac3ba7485eb92e23a2ce2cd1a1cca34c32"} Oct 08 00:08:55 crc kubenswrapper[4967]: I1008 00:08:55.311169 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f3104984097ec392438cf7eefed160ac3ba7485eb92e23a2ce2cd1a1cca34c32" Oct 08 00:08:55 crc kubenswrapper[4967]: I1008 00:08:55.311258 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 08 00:08:55 crc kubenswrapper[4967]: I1008 00:08:55.316648 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"85a3da15-9895-4e7f-81f4-6be5e34061de","Type":"ContainerStarted","Data":"21cf9e01b29c788b8fb94fdb2820b913be7ccc1009e0860c72bf6548caa84329"} Oct 08 00:08:55 crc kubenswrapper[4967]: I1008 00:08:55.390815 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-p9kxf" Oct 08 00:08:55 crc kubenswrapper[4967]: I1008 00:08:55.402236 4967 patch_prober.go:28] interesting pod/router-default-5444994796-p9kxf container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 00:08:55 crc kubenswrapper[4967]: [-]has-synced failed: reason withheld Oct 08 00:08:55 crc kubenswrapper[4967]: [+]process-running ok Oct 08 00:08:55 crc kubenswrapper[4967]: healthz check failed Oct 08 00:08:55 crc kubenswrapper[4967]: I1008 00:08:55.402313 4967 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-p9kxf" podUID="ee7dadb3-3ab7-4790-9b67-7f7a5bf7604f" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 00:08:56 crc kubenswrapper[4967]: I1008 00:08:56.337607 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"85a3da15-9895-4e7f-81f4-6be5e34061de","Type":"ContainerStarted","Data":"d3c297d6f8f8d31cbcca4d2722766ffe53c1271f2ddd1006a3a8a609b620b002"} Oct 08 00:08:56 crc kubenswrapper[4967]: I1008 00:08:56.350074 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=3.35005746 podStartE2EDuration="3.35005746s" podCreationTimestamp="2025-10-08 00:08:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:08:56.349502162 +0000 UTC m=+135.000036811" watchObservedRunningTime="2025-10-08 00:08:56.35005746 +0000 UTC m=+135.000592109" Oct 08 00:08:56 crc kubenswrapper[4967]: I1008 00:08:56.393241 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-p9kxf" Oct 08 00:08:56 crc kubenswrapper[4967]: I1008 00:08:56.394916 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-p9kxf" Oct 08 00:08:57 crc kubenswrapper[4967]: I1008 00:08:57.357334 4967 generic.go:334] "Generic (PLEG): container finished" podID="85a3da15-9895-4e7f-81f4-6be5e34061de" containerID="d3c297d6f8f8d31cbcca4d2722766ffe53c1271f2ddd1006a3a8a609b620b002" exitCode=0 Oct 08 00:08:57 crc kubenswrapper[4967]: I1008 00:08:57.357431 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"85a3da15-9895-4e7f-81f4-6be5e34061de","Type":"ContainerDied","Data":"d3c297d6f8f8d31cbcca4d2722766ffe53c1271f2ddd1006a3a8a609b620b002"} Oct 08 00:08:59 crc kubenswrapper[4967]: I1008 00:08:59.641842 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-fjz5z" Oct 08 00:08:59 crc kubenswrapper[4967]: I1008 00:08:59.900740 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" Oct 08 00:09:01 crc kubenswrapper[4967]: I1008 00:09:01.977192 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-tdqfv" Oct 08 00:09:04 crc kubenswrapper[4967]: I1008 00:09:04.719600 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-vxdxg" Oct 08 00:09:04 crc kubenswrapper[4967]: I1008 00:09:04.725196 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-vxdxg" Oct 08 00:09:04 crc kubenswrapper[4967]: I1008 00:09:04.901359 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 08 00:09:05 crc kubenswrapper[4967]: I1008 00:09:05.032402 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/85a3da15-9895-4e7f-81f4-6be5e34061de-kubelet-dir\") pod \"85a3da15-9895-4e7f-81f4-6be5e34061de\" (UID: \"85a3da15-9895-4e7f-81f4-6be5e34061de\") " Oct 08 00:09:05 crc kubenswrapper[4967]: I1008 00:09:05.032871 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/85a3da15-9895-4e7f-81f4-6be5e34061de-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "85a3da15-9895-4e7f-81f4-6be5e34061de" (UID: "85a3da15-9895-4e7f-81f4-6be5e34061de"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 00:09:05 crc kubenswrapper[4967]: I1008 00:09:05.033077 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/85a3da15-9895-4e7f-81f4-6be5e34061de-kube-api-access\") pod \"85a3da15-9895-4e7f-81f4-6be5e34061de\" (UID: \"85a3da15-9895-4e7f-81f4-6be5e34061de\") " Oct 08 00:09:05 crc kubenswrapper[4967]: I1008 00:09:05.033479 4967 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/85a3da15-9895-4e7f-81f4-6be5e34061de-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 08 00:09:05 crc kubenswrapper[4967]: I1008 00:09:05.052909 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85a3da15-9895-4e7f-81f4-6be5e34061de-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "85a3da15-9895-4e7f-81f4-6be5e34061de" (UID: "85a3da15-9895-4e7f-81f4-6be5e34061de"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:09:05 crc kubenswrapper[4967]: I1008 00:09:05.136886 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/85a3da15-9895-4e7f-81f4-6be5e34061de-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 08 00:09:05 crc kubenswrapper[4967]: I1008 00:09:05.518471 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 08 00:09:05 crc kubenswrapper[4967]: I1008 00:09:05.525796 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"85a3da15-9895-4e7f-81f4-6be5e34061de","Type":"ContainerDied","Data":"21cf9e01b29c788b8fb94fdb2820b913be7ccc1009e0860c72bf6548caa84329"} Oct 08 00:09:05 crc kubenswrapper[4967]: I1008 00:09:05.525825 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="21cf9e01b29c788b8fb94fdb2820b913be7ccc1009e0860c72bf6548caa84329" Oct 08 00:09:06 crc kubenswrapper[4967]: I1008 00:09:06.920625 4967 patch_prober.go:28] interesting pod/machine-config-daemon-xvdtp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 00:09:06 crc kubenswrapper[4967]: I1008 00:09:06.921029 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" podUID="9461162a-c27d-465e-9ca7-2d3936877cd8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 00:09:11 crc kubenswrapper[4967]: I1008 00:09:11.137347 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 00:09:11 crc kubenswrapper[4967]: I1008 00:09:11.137941 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 00:09:11 crc kubenswrapper[4967]: I1008 00:09:11.137985 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 00:09:11 crc kubenswrapper[4967]: I1008 00:09:11.138017 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 00:09:11 crc kubenswrapper[4967]: I1008 00:09:11.139086 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 00:09:11 crc kubenswrapper[4967]: I1008 00:09:11.142401 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 00:09:11 crc kubenswrapper[4967]: I1008 00:09:11.142698 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 00:09:11 crc kubenswrapper[4967]: I1008 00:09:11.145476 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 00:09:11 crc kubenswrapper[4967]: I1008 00:09:11.146724 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 00:09:11 crc kubenswrapper[4967]: I1008 00:09:11.171014 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 00:09:11 crc kubenswrapper[4967]: I1008 00:09:11.372629 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 00:09:11 crc kubenswrapper[4967]: I1008 00:09:11.923284 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:09:18 crc kubenswrapper[4967]: E1008 00:09:18.577446 4967 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 08 00:09:18 crc kubenswrapper[4967]: E1008 00:09:18.578259 4967 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-t46zc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-9sp7f_openshift-marketplace(a0f7251c-fe97-4f8d-8121-82f7010170ef): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 08 00:09:18 crc kubenswrapper[4967]: E1008 00:09:18.579475 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-9sp7f" podUID="a0f7251c-fe97-4f8d-8121-82f7010170ef" Oct 08 00:09:18 crc kubenswrapper[4967]: I1008 00:09:18.610564 4967 generic.go:334] "Generic (PLEG): container finished" podID="84cd9a7a-bb01-43de-9055-deccbc8529b2" containerID="8a3b4d3efcc161d3a54910bdf7e3aaf8f349acedd7bdf581c2b31f74eed0f3ae" exitCode=0 Oct 08 00:09:18 crc kubenswrapper[4967]: I1008 00:09:18.610658 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-pruner-29331360-494rn" event={"ID":"84cd9a7a-bb01-43de-9055-deccbc8529b2","Type":"ContainerDied","Data":"8a3b4d3efcc161d3a54910bdf7e3aaf8f349acedd7bdf581c2b31f74eed0f3ae"} Oct 08 00:09:19 crc kubenswrapper[4967]: E1008 00:09:19.849981 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-9sp7f" podUID="a0f7251c-fe97-4f8d-8121-82f7010170ef" Oct 08 00:09:19 crc kubenswrapper[4967]: E1008 00:09:19.940902 4967 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 08 00:09:19 crc kubenswrapper[4967]: E1008 00:09:19.941057 4967 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-775qd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-x8d4p_openshift-marketplace(32d2d299-2970-40e9-9116-74a074c80eb1): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 08 00:09:19 crc kubenswrapper[4967]: E1008 00:09:19.942519 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-x8d4p" podUID="32d2d299-2970-40e9-9116-74a074c80eb1" Oct 08 00:09:21 crc kubenswrapper[4967]: E1008 00:09:21.191412 4967 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 08 00:09:21 crc kubenswrapper[4967]: E1008 00:09:21.191900 4967 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zq8z8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-wcll6_openshift-marketplace(9c827bbc-b90c-4168-918d-c2658da9b6ac): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 08 00:09:21 crc kubenswrapper[4967]: E1008 00:09:21.193030 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-wcll6" podUID="9c827bbc-b90c-4168-918d-c2658da9b6ac" Oct 08 00:09:21 crc kubenswrapper[4967]: E1008 00:09:21.202313 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-x8d4p" podUID="32d2d299-2970-40e9-9116-74a074c80eb1" Oct 08 00:09:21 crc kubenswrapper[4967]: E1008 00:09:21.267570 4967 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 08 00:09:21 crc kubenswrapper[4967]: E1008 00:09:21.267740 4967 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-blhnw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-9glfc_openshift-marketplace(f174780f-4c64-427f-b527-91a3871c271d): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 08 00:09:21 crc kubenswrapper[4967]: E1008 00:09:21.268942 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-9glfc" podUID="f174780f-4c64-427f-b527-91a3871c271d" Oct 08 00:09:21 crc kubenswrapper[4967]: I1008 00:09:21.291329 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-pruner-29331360-494rn" Oct 08 00:09:21 crc kubenswrapper[4967]: I1008 00:09:21.435088 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/84cd9a7a-bb01-43de-9055-deccbc8529b2-serviceca\") pod \"84cd9a7a-bb01-43de-9055-deccbc8529b2\" (UID: \"84cd9a7a-bb01-43de-9055-deccbc8529b2\") " Oct 08 00:09:21 crc kubenswrapper[4967]: I1008 00:09:21.435213 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t2th5\" (UniqueName: \"kubernetes.io/projected/84cd9a7a-bb01-43de-9055-deccbc8529b2-kube-api-access-t2th5\") pod \"84cd9a7a-bb01-43de-9055-deccbc8529b2\" (UID: \"84cd9a7a-bb01-43de-9055-deccbc8529b2\") " Oct 08 00:09:21 crc kubenswrapper[4967]: I1008 00:09:21.436405 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/84cd9a7a-bb01-43de-9055-deccbc8529b2-serviceca" (OuterVolumeSpecName: "serviceca") pod "84cd9a7a-bb01-43de-9055-deccbc8529b2" (UID: "84cd9a7a-bb01-43de-9055-deccbc8529b2"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:09:21 crc kubenswrapper[4967]: I1008 00:09:21.441131 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84cd9a7a-bb01-43de-9055-deccbc8529b2-kube-api-access-t2th5" (OuterVolumeSpecName: "kube-api-access-t2th5") pod "84cd9a7a-bb01-43de-9055-deccbc8529b2" (UID: "84cd9a7a-bb01-43de-9055-deccbc8529b2"). InnerVolumeSpecName "kube-api-access-t2th5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:09:21 crc kubenswrapper[4967]: I1008 00:09:21.536510 4967 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/84cd9a7a-bb01-43de-9055-deccbc8529b2-serviceca\") on node \"crc\" DevicePath \"\"" Oct 08 00:09:21 crc kubenswrapper[4967]: I1008 00:09:21.536546 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t2th5\" (UniqueName: \"kubernetes.io/projected/84cd9a7a-bb01-43de-9055-deccbc8529b2-kube-api-access-t2th5\") on node \"crc\" DevicePath \"\"" Oct 08 00:09:21 crc kubenswrapper[4967]: I1008 00:09:21.628378 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-pruner-29331360-494rn" event={"ID":"84cd9a7a-bb01-43de-9055-deccbc8529b2","Type":"ContainerDied","Data":"20c016534c059bad4d1f03867754c83e11b1d339f8550a6b47852de1dae5f9da"} Oct 08 00:09:21 crc kubenswrapper[4967]: I1008 00:09:21.628432 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="20c016534c059bad4d1f03867754c83e11b1d339f8550a6b47852de1dae5f9da" Oct 08 00:09:21 crc kubenswrapper[4967]: I1008 00:09:21.628527 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-pruner-29331360-494rn" Oct 08 00:09:24 crc kubenswrapper[4967]: I1008 00:09:24.124267 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b5glr" Oct 08 00:09:24 crc kubenswrapper[4967]: E1008 00:09:24.266116 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-9glfc" podUID="f174780f-4c64-427f-b527-91a3871c271d" Oct 08 00:09:24 crc kubenswrapper[4967]: E1008 00:09:24.266165 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-wcll6" podUID="9c827bbc-b90c-4168-918d-c2658da9b6ac" Oct 08 00:09:24 crc kubenswrapper[4967]: E1008 00:09:24.344489 4967 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Oct 08 00:09:24 crc kubenswrapper[4967]: E1008 00:09:24.344703 4967 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5qq8z,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-v7scw_openshift-marketplace(f3001d43-c530-4c5e-a6e6-27346682ba5c): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 08 00:09:24 crc kubenswrapper[4967]: E1008 00:09:24.345801 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-v7scw" podUID="f3001d43-c530-4c5e-a6e6-27346682ba5c" Oct 08 00:09:24 crc kubenswrapper[4967]: E1008 00:09:24.353806 4967 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 08 00:09:24 crc kubenswrapper[4967]: E1008 00:09:24.353940 4967 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fwcx7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-4hhcd_openshift-marketplace(8912b04d-fd70-497c-aab6-7490fc6ddf40): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 08 00:09:24 crc kubenswrapper[4967]: E1008 00:09:24.355612 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-4hhcd" podUID="8912b04d-fd70-497c-aab6-7490fc6ddf40" Oct 08 00:09:24 crc kubenswrapper[4967]: E1008 00:09:24.439994 4967 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Oct 08 00:09:24 crc kubenswrapper[4967]: E1008 00:09:24.441190 4967 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-nf2kn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-f24p2_openshift-marketplace(92a4f64f-e1cb-42a1-85b4-d3fc38e949a2): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 08 00:09:24 crc kubenswrapper[4967]: E1008 00:09:24.442379 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-f24p2" podUID="92a4f64f-e1cb-42a1-85b4-d3fc38e949a2" Oct 08 00:09:24 crc kubenswrapper[4967]: W1008 00:09:24.484742 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-e5d9e14ee52f52133a5a9524e91480bc4e27e0a5626d3f677ac1e89182a059b9 WatchSource:0}: Error finding container e5d9e14ee52f52133a5a9524e91480bc4e27e0a5626d3f677ac1e89182a059b9: Status 404 returned error can't find the container with id e5d9e14ee52f52133a5a9524e91480bc4e27e0a5626d3f677ac1e89182a059b9 Oct 08 00:09:24 crc kubenswrapper[4967]: I1008 00:09:24.646142 4967 generic.go:334] "Generic (PLEG): container finished" podID="f0340c31-098f-4936-bd5e-4ef17c48a9f1" containerID="22507f8b5ef6f57835f248285e841020cdaf6f9a17f62e5ec6a75d3bdeaaa127" exitCode=0 Oct 08 00:09:24 crc kubenswrapper[4967]: I1008 00:09:24.646222 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rzc76" event={"ID":"f0340c31-098f-4936-bd5e-4ef17c48a9f1","Type":"ContainerDied","Data":"22507f8b5ef6f57835f248285e841020cdaf6f9a17f62e5ec6a75d3bdeaaa127"} Oct 08 00:09:24 crc kubenswrapper[4967]: I1008 00:09:24.652215 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"d830493f4a67c3610b8adc97cf886bb37780647e71898e535fb5e118b2ff2fe2"} Oct 08 00:09:24 crc kubenswrapper[4967]: I1008 00:09:24.652254 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"e5d9e14ee52f52133a5a9524e91480bc4e27e0a5626d3f677ac1e89182a059b9"} Oct 08 00:09:24 crc kubenswrapper[4967]: E1008 00:09:24.652450 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-v7scw" podUID="f3001d43-c530-4c5e-a6e6-27346682ba5c" Oct 08 00:09:24 crc kubenswrapper[4967]: E1008 00:09:24.652816 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-f24p2" podUID="92a4f64f-e1cb-42a1-85b4-d3fc38e949a2" Oct 08 00:09:24 crc kubenswrapper[4967]: E1008 00:09:24.652834 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-4hhcd" podUID="8912b04d-fd70-497c-aab6-7490fc6ddf40" Oct 08 00:09:24 crc kubenswrapper[4967]: W1008 00:09:24.763418 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-2eb426e9f64fb01f62338e9b92e7e9f980d520b36d0cd0aeb2b858800e282182 WatchSource:0}: Error finding container 2eb426e9f64fb01f62338e9b92e7e9f980d520b36d0cd0aeb2b858800e282182: Status 404 returned error can't find the container with id 2eb426e9f64fb01f62338e9b92e7e9f980d520b36d0cd0aeb2b858800e282182 Oct 08 00:09:25 crc kubenswrapper[4967]: I1008 00:09:25.661425 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rzc76" event={"ID":"f0340c31-098f-4936-bd5e-4ef17c48a9f1","Type":"ContainerStarted","Data":"096b92263447d7226b4eaf48f8590d960262063ffc31549f9efc25a3eacd8a64"} Oct 08 00:09:25 crc kubenswrapper[4967]: I1008 00:09:25.663631 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"92e32b33419f444654afc667a9aedb727e7bdb04a602a0459d2b319aaab4f694"} Oct 08 00:09:25 crc kubenswrapper[4967]: I1008 00:09:25.663739 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"2eb426e9f64fb01f62338e9b92e7e9f980d520b36d0cd0aeb2b858800e282182"} Oct 08 00:09:25 crc kubenswrapper[4967]: I1008 00:09:25.666914 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"544b1d1463da76fc11de221a451919a95ded479a183d9b76fb37094cc0418c91"} Oct 08 00:09:25 crc kubenswrapper[4967]: I1008 00:09:25.666976 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"c051b1cbdddf1a4538992b5ebb9d876b9ef2116f35348dfafb1f2b998ee438d0"} Oct 08 00:09:25 crc kubenswrapper[4967]: I1008 00:09:25.667258 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 00:09:25 crc kubenswrapper[4967]: I1008 00:09:25.693184 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-rzc76" podStartSLOduration=2.83299407 podStartE2EDuration="35.693163159s" podCreationTimestamp="2025-10-08 00:08:50 +0000 UTC" firstStartedPulling="2025-10-08 00:08:52.190889 +0000 UTC m=+130.841423649" lastFinishedPulling="2025-10-08 00:09:25.051058089 +0000 UTC m=+163.701592738" observedRunningTime="2025-10-08 00:09:25.689814793 +0000 UTC m=+164.340349472" watchObservedRunningTime="2025-10-08 00:09:25.693163159 +0000 UTC m=+164.343697808" Oct 08 00:09:27 crc kubenswrapper[4967]: I1008 00:09:27.005799 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/59f68771-e25d-4367-87ee-68a31d767882-metrics-certs\") pod \"network-metrics-daemon-2h7z8\" (UID: \"59f68771-e25d-4367-87ee-68a31d767882\") " pod="openshift-multus/network-metrics-daemon-2h7z8" Oct 08 00:09:27 crc kubenswrapper[4967]: I1008 00:09:27.014324 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/59f68771-e25d-4367-87ee-68a31d767882-metrics-certs\") pod \"network-metrics-daemon-2h7z8\" (UID: \"59f68771-e25d-4367-87ee-68a31d767882\") " pod="openshift-multus/network-metrics-daemon-2h7z8" Oct 08 00:09:27 crc kubenswrapper[4967]: I1008 00:09:27.063512 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-2h7z8" Oct 08 00:09:27 crc kubenswrapper[4967]: I1008 00:09:27.504191 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-2h7z8"] Oct 08 00:09:27 crc kubenswrapper[4967]: W1008 00:09:27.521422 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod59f68771_e25d_4367_87ee_68a31d767882.slice/crio-9887619c8d7e3ce3940d74795f0b5268c0637b53868b600bc586e7e82a3e0d7a WatchSource:0}: Error finding container 9887619c8d7e3ce3940d74795f0b5268c0637b53868b600bc586e7e82a3e0d7a: Status 404 returned error can't find the container with id 9887619c8d7e3ce3940d74795f0b5268c0637b53868b600bc586e7e82a3e0d7a Oct 08 00:09:27 crc kubenswrapper[4967]: I1008 00:09:27.679130 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-2h7z8" event={"ID":"59f68771-e25d-4367-87ee-68a31d767882","Type":"ContainerStarted","Data":"9887619c8d7e3ce3940d74795f0b5268c0637b53868b600bc586e7e82a3e0d7a"} Oct 08 00:09:28 crc kubenswrapper[4967]: I1008 00:09:28.692579 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-2h7z8" event={"ID":"59f68771-e25d-4367-87ee-68a31d767882","Type":"ContainerStarted","Data":"74ddc7806ff07c90fd0c8c4858c46ee382357baa40438e1de9d449587dc76640"} Oct 08 00:09:28 crc kubenswrapper[4967]: I1008 00:09:28.692934 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-2h7z8" event={"ID":"59f68771-e25d-4367-87ee-68a31d767882","Type":"ContainerStarted","Data":"9ab6c482e10894c92d5399c0aa50dc97de5aa38baa7d374dce9d5329660d4032"} Oct 08 00:09:28 crc kubenswrapper[4967]: I1008 00:09:28.713602 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-2h7z8" podStartSLOduration=145.713573586 podStartE2EDuration="2m25.713573586s" podCreationTimestamp="2025-10-08 00:07:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:09:28.710654413 +0000 UTC m=+167.361189102" watchObservedRunningTime="2025-10-08 00:09:28.713573586 +0000 UTC m=+167.364108265" Oct 08 00:09:31 crc kubenswrapper[4967]: I1008 00:09:31.002132 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-rzc76" Oct 08 00:09:31 crc kubenswrapper[4967]: I1008 00:09:31.002528 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-rzc76" Oct 08 00:09:31 crc kubenswrapper[4967]: I1008 00:09:31.173497 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-rzc76" Oct 08 00:09:31 crc kubenswrapper[4967]: I1008 00:09:31.761578 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-rzc76" Oct 08 00:09:32 crc kubenswrapper[4967]: I1008 00:09:32.722032 4967 generic.go:334] "Generic (PLEG): container finished" podID="a0f7251c-fe97-4f8d-8121-82f7010170ef" containerID="6bbc46d8a8a19157a3c16a5fa630506dbaa7d6488a3cc69dc401687eca3d7b63" exitCode=0 Oct 08 00:09:32 crc kubenswrapper[4967]: I1008 00:09:32.723428 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9sp7f" event={"ID":"a0f7251c-fe97-4f8d-8121-82f7010170ef","Type":"ContainerDied","Data":"6bbc46d8a8a19157a3c16a5fa630506dbaa7d6488a3cc69dc401687eca3d7b63"} Oct 08 00:09:33 crc kubenswrapper[4967]: I1008 00:09:33.730835 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9sp7f" event={"ID":"a0f7251c-fe97-4f8d-8121-82f7010170ef","Type":"ContainerStarted","Data":"b9d92067efb26616bac31ab90d833e2ef6f3b8f02c1cc629588232e2a67ce396"} Oct 08 00:09:33 crc kubenswrapper[4967]: I1008 00:09:33.758346 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-9sp7f" podStartSLOduration=3.654976122 podStartE2EDuration="45.758328067s" podCreationTimestamp="2025-10-08 00:08:48 +0000 UTC" firstStartedPulling="2025-10-08 00:08:51.069780744 +0000 UTC m=+129.720315393" lastFinishedPulling="2025-10-08 00:09:33.173132649 +0000 UTC m=+171.823667338" observedRunningTime="2025-10-08 00:09:33.757396748 +0000 UTC m=+172.407931437" watchObservedRunningTime="2025-10-08 00:09:33.758328067 +0000 UTC m=+172.408862736" Oct 08 00:09:34 crc kubenswrapper[4967]: I1008 00:09:34.739517 4967 generic.go:334] "Generic (PLEG): container finished" podID="32d2d299-2970-40e9-9116-74a074c80eb1" containerID="1430349593ed2e61050f2eb745cfad0b82ef9f6e459baee55b2d54a8c2c11593" exitCode=0 Oct 08 00:09:34 crc kubenswrapper[4967]: I1008 00:09:34.739567 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x8d4p" event={"ID":"32d2d299-2970-40e9-9116-74a074c80eb1","Type":"ContainerDied","Data":"1430349593ed2e61050f2eb745cfad0b82ef9f6e459baee55b2d54a8c2c11593"} Oct 08 00:09:35 crc kubenswrapper[4967]: I1008 00:09:35.749616 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x8d4p" event={"ID":"32d2d299-2970-40e9-9116-74a074c80eb1","Type":"ContainerStarted","Data":"1a3ee19601a511037afb82db3574b9dbfd14a0521374a3a7993a076bbfd08ea7"} Oct 08 00:09:36 crc kubenswrapper[4967]: I1008 00:09:36.758816 4967 generic.go:334] "Generic (PLEG): container finished" podID="9c827bbc-b90c-4168-918d-c2658da9b6ac" containerID="a5db13822d13533dfbfddedd2e552ec1030b1ce84abd8a1c3250e83cabdbb5c1" exitCode=0 Oct 08 00:09:36 crc kubenswrapper[4967]: I1008 00:09:36.758865 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wcll6" event={"ID":"9c827bbc-b90c-4168-918d-c2658da9b6ac","Type":"ContainerDied","Data":"a5db13822d13533dfbfddedd2e552ec1030b1ce84abd8a1c3250e83cabdbb5c1"} Oct 08 00:09:36 crc kubenswrapper[4967]: I1008 00:09:36.781515 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-x8d4p" podStartSLOduration=4.71235722 podStartE2EDuration="48.781491962s" podCreationTimestamp="2025-10-08 00:08:48 +0000 UTC" firstStartedPulling="2025-10-08 00:08:51.085058842 +0000 UTC m=+129.735593491" lastFinishedPulling="2025-10-08 00:09:35.154193594 +0000 UTC m=+173.804728233" observedRunningTime="2025-10-08 00:09:35.777108853 +0000 UTC m=+174.427643513" watchObservedRunningTime="2025-10-08 00:09:36.781491962 +0000 UTC m=+175.432026611" Oct 08 00:09:36 crc kubenswrapper[4967]: I1008 00:09:36.920194 4967 patch_prober.go:28] interesting pod/machine-config-daemon-xvdtp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 00:09:36 crc kubenswrapper[4967]: I1008 00:09:36.920472 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" podUID="9461162a-c27d-465e-9ca7-2d3936877cd8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 00:09:37 crc kubenswrapper[4967]: I1008 00:09:37.767150 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wcll6" event={"ID":"9c827bbc-b90c-4168-918d-c2658da9b6ac","Type":"ContainerStarted","Data":"c0e93445fdd2997689d7ad3f47438f178238356a97d3da6986d3c0fd71645ea0"} Oct 08 00:09:37 crc kubenswrapper[4967]: I1008 00:09:37.795481 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-wcll6" podStartSLOduration=1.833246994 podStartE2EDuration="46.795460095s" podCreationTimestamp="2025-10-08 00:08:51 +0000 UTC" firstStartedPulling="2025-10-08 00:08:52.23685651 +0000 UTC m=+130.887391159" lastFinishedPulling="2025-10-08 00:09:37.199069601 +0000 UTC m=+175.849604260" observedRunningTime="2025-10-08 00:09:37.792617415 +0000 UTC m=+176.443152074" watchObservedRunningTime="2025-10-08 00:09:37.795460095 +0000 UTC m=+176.445994754" Oct 08 00:09:39 crc kubenswrapper[4967]: I1008 00:09:39.342579 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-x8d4p" Oct 08 00:09:39 crc kubenswrapper[4967]: I1008 00:09:39.342948 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-x8d4p" Oct 08 00:09:39 crc kubenswrapper[4967]: I1008 00:09:39.349422 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-9sp7f" Oct 08 00:09:39 crc kubenswrapper[4967]: I1008 00:09:39.349483 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-9sp7f" Oct 08 00:09:39 crc kubenswrapper[4967]: I1008 00:09:39.409220 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-x8d4p" Oct 08 00:09:39 crc kubenswrapper[4967]: I1008 00:09:39.418411 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-9sp7f" Oct 08 00:09:39 crc kubenswrapper[4967]: I1008 00:09:39.779991 4967 generic.go:334] "Generic (PLEG): container finished" podID="8912b04d-fd70-497c-aab6-7490fc6ddf40" containerID="e2f7ac3d2d6f7a8672da3676f8bc923f89da5b022af990fe7f3d8be5b95a2743" exitCode=0 Oct 08 00:09:39 crc kubenswrapper[4967]: I1008 00:09:39.780089 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4hhcd" event={"ID":"8912b04d-fd70-497c-aab6-7490fc6ddf40","Type":"ContainerDied","Data":"e2f7ac3d2d6f7a8672da3676f8bc923f89da5b022af990fe7f3d8be5b95a2743"} Oct 08 00:09:39 crc kubenswrapper[4967]: I1008 00:09:39.783057 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v7scw" event={"ID":"f3001d43-c530-4c5e-a6e6-27346682ba5c","Type":"ContainerStarted","Data":"454fcde55a49691a74ac4ad160859f26321f0cf958740208ac85065d2a653837"} Oct 08 00:09:39 crc kubenswrapper[4967]: I1008 00:09:39.833090 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-9sp7f" Oct 08 00:09:40 crc kubenswrapper[4967]: I1008 00:09:40.790617 4967 generic.go:334] "Generic (PLEG): container finished" podID="f174780f-4c64-427f-b527-91a3871c271d" containerID="f016083cd27bea7cf6bf1f296c6f9468ea1645fffa309886dcf8a3e9bfd1e5df" exitCode=0 Oct 08 00:09:40 crc kubenswrapper[4967]: I1008 00:09:40.790715 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9glfc" event={"ID":"f174780f-4c64-427f-b527-91a3871c271d","Type":"ContainerDied","Data":"f016083cd27bea7cf6bf1f296c6f9468ea1645fffa309886dcf8a3e9bfd1e5df"} Oct 08 00:09:40 crc kubenswrapper[4967]: I1008 00:09:40.794309 4967 generic.go:334] "Generic (PLEG): container finished" podID="f3001d43-c530-4c5e-a6e6-27346682ba5c" containerID="454fcde55a49691a74ac4ad160859f26321f0cf958740208ac85065d2a653837" exitCode=0 Oct 08 00:09:40 crc kubenswrapper[4967]: I1008 00:09:40.794392 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v7scw" event={"ID":"f3001d43-c530-4c5e-a6e6-27346682ba5c","Type":"ContainerDied","Data":"454fcde55a49691a74ac4ad160859f26321f0cf958740208ac85065d2a653837"} Oct 08 00:09:40 crc kubenswrapper[4967]: I1008 00:09:40.803749 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4hhcd" event={"ID":"8912b04d-fd70-497c-aab6-7490fc6ddf40","Type":"ContainerStarted","Data":"e7db1c6944d8b20adde1074eead4d289d796a9369b0a8ed3736e49af93449098"} Oct 08 00:09:40 crc kubenswrapper[4967]: I1008 00:09:40.855728 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-4hhcd" podStartSLOduration=2.823904246 podStartE2EDuration="51.855711982s" podCreationTimestamp="2025-10-08 00:08:49 +0000 UTC" firstStartedPulling="2025-10-08 00:08:51.169897956 +0000 UTC m=+129.820432605" lastFinishedPulling="2025-10-08 00:09:40.201705682 +0000 UTC m=+178.852240341" observedRunningTime="2025-10-08 00:09:40.85408381 +0000 UTC m=+179.504618489" watchObservedRunningTime="2025-10-08 00:09:40.855711982 +0000 UTC m=+179.506246631" Oct 08 00:09:41 crc kubenswrapper[4967]: I1008 00:09:41.420428 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-wcll6" Oct 08 00:09:41 crc kubenswrapper[4967]: I1008 00:09:41.420900 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-wcll6" Oct 08 00:09:41 crc kubenswrapper[4967]: I1008 00:09:41.466203 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-wcll6" Oct 08 00:09:41 crc kubenswrapper[4967]: I1008 00:09:41.811753 4967 generic.go:334] "Generic (PLEG): container finished" podID="92a4f64f-e1cb-42a1-85b4-d3fc38e949a2" containerID="d013ec6e24839f4a89cbc8aea45a44ca43c39887969b94e9b9e0e452fceace04" exitCode=0 Oct 08 00:09:41 crc kubenswrapper[4967]: I1008 00:09:41.812583 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f24p2" event={"ID":"92a4f64f-e1cb-42a1-85b4-d3fc38e949a2","Type":"ContainerDied","Data":"d013ec6e24839f4a89cbc8aea45a44ca43c39887969b94e9b9e0e452fceace04"} Oct 08 00:09:42 crc kubenswrapper[4967]: I1008 00:09:42.819331 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f24p2" event={"ID":"92a4f64f-e1cb-42a1-85b4-d3fc38e949a2","Type":"ContainerStarted","Data":"049fc4c8ddf143ec36fee0da8c5dd5c26592cca956aa969e73472747ab22e966"} Oct 08 00:09:42 crc kubenswrapper[4967]: I1008 00:09:42.822494 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v7scw" event={"ID":"f3001d43-c530-4c5e-a6e6-27346682ba5c","Type":"ContainerStarted","Data":"4ec9fbb13bbd2a22a32c0a838bf811148fdc82af7c822a8b9a9073d9c1baf64e"} Oct 08 00:09:42 crc kubenswrapper[4967]: I1008 00:09:42.824542 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9glfc" event={"ID":"f174780f-4c64-427f-b527-91a3871c271d","Type":"ContainerStarted","Data":"015cf08ab330b3b2ce10e6a71da3dcae87b198a098e916db2599b8b24d67437c"} Oct 08 00:09:42 crc kubenswrapper[4967]: I1008 00:09:42.839188 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-f24p2" podStartSLOduration=1.795181835 podStartE2EDuration="50.839166743s" podCreationTimestamp="2025-10-08 00:08:52 +0000 UTC" firstStartedPulling="2025-10-08 00:08:53.271186506 +0000 UTC m=+131.921721155" lastFinishedPulling="2025-10-08 00:09:42.315171414 +0000 UTC m=+180.965706063" observedRunningTime="2025-10-08 00:09:42.837120438 +0000 UTC m=+181.487655097" watchObservedRunningTime="2025-10-08 00:09:42.839166743 +0000 UTC m=+181.489701412" Oct 08 00:09:42 crc kubenswrapper[4967]: I1008 00:09:42.854489 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-9glfc" podStartSLOduration=3.217189364 podStartE2EDuration="54.85447156s" podCreationTimestamp="2025-10-08 00:08:48 +0000 UTC" firstStartedPulling="2025-10-08 00:08:50.018891589 +0000 UTC m=+128.669426238" lastFinishedPulling="2025-10-08 00:09:41.656173785 +0000 UTC m=+180.306708434" observedRunningTime="2025-10-08 00:09:42.850519895 +0000 UTC m=+181.501054544" watchObservedRunningTime="2025-10-08 00:09:42.85447156 +0000 UTC m=+181.505006209" Oct 08 00:09:42 crc kubenswrapper[4967]: I1008 00:09:42.871954 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-v7scw" podStartSLOduration=3.59671391 podStartE2EDuration="51.871928346s" podCreationTimestamp="2025-10-08 00:08:51 +0000 UTC" firstStartedPulling="2025-10-08 00:08:53.271520544 +0000 UTC m=+131.922055183" lastFinishedPulling="2025-10-08 00:09:41.54673497 +0000 UTC m=+180.197269619" observedRunningTime="2025-10-08 00:09:42.866681469 +0000 UTC m=+181.517216128" watchObservedRunningTime="2025-10-08 00:09:42.871928346 +0000 UTC m=+181.522463005" Oct 08 00:09:43 crc kubenswrapper[4967]: I1008 00:09:43.554239 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9sp7f"] Oct 08 00:09:43 crc kubenswrapper[4967]: I1008 00:09:43.555077 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-9sp7f" podUID="a0f7251c-fe97-4f8d-8121-82f7010170ef" containerName="registry-server" containerID="cri-o://b9d92067efb26616bac31ab90d833e2ef6f3b8f02c1cc629588232e2a67ce396" gracePeriod=2 Oct 08 00:09:43 crc kubenswrapper[4967]: E1008 00:09:43.726826 4967 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda0f7251c_fe97_4f8d_8121_82f7010170ef.slice/crio-b9d92067efb26616bac31ab90d833e2ef6f3b8f02c1cc629588232e2a67ce396.scope\": RecentStats: unable to find data in memory cache]" Oct 08 00:09:44 crc kubenswrapper[4967]: I1008 00:09:44.407656 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9sp7f" Oct 08 00:09:44 crc kubenswrapper[4967]: I1008 00:09:44.545038 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0f7251c-fe97-4f8d-8121-82f7010170ef-catalog-content\") pod \"a0f7251c-fe97-4f8d-8121-82f7010170ef\" (UID: \"a0f7251c-fe97-4f8d-8121-82f7010170ef\") " Oct 08 00:09:44 crc kubenswrapper[4967]: I1008 00:09:44.545429 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0f7251c-fe97-4f8d-8121-82f7010170ef-utilities\") pod \"a0f7251c-fe97-4f8d-8121-82f7010170ef\" (UID: \"a0f7251c-fe97-4f8d-8121-82f7010170ef\") " Oct 08 00:09:44 crc kubenswrapper[4967]: I1008 00:09:44.545453 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t46zc\" (UniqueName: \"kubernetes.io/projected/a0f7251c-fe97-4f8d-8121-82f7010170ef-kube-api-access-t46zc\") pod \"a0f7251c-fe97-4f8d-8121-82f7010170ef\" (UID: \"a0f7251c-fe97-4f8d-8121-82f7010170ef\") " Oct 08 00:09:44 crc kubenswrapper[4967]: I1008 00:09:44.546176 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a0f7251c-fe97-4f8d-8121-82f7010170ef-utilities" (OuterVolumeSpecName: "utilities") pod "a0f7251c-fe97-4f8d-8121-82f7010170ef" (UID: "a0f7251c-fe97-4f8d-8121-82f7010170ef"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:09:44 crc kubenswrapper[4967]: I1008 00:09:44.550831 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0f7251c-fe97-4f8d-8121-82f7010170ef-kube-api-access-t46zc" (OuterVolumeSpecName: "kube-api-access-t46zc") pod "a0f7251c-fe97-4f8d-8121-82f7010170ef" (UID: "a0f7251c-fe97-4f8d-8121-82f7010170ef"). InnerVolumeSpecName "kube-api-access-t46zc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:09:44 crc kubenswrapper[4967]: I1008 00:09:44.591945 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a0f7251c-fe97-4f8d-8121-82f7010170ef-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a0f7251c-fe97-4f8d-8121-82f7010170ef" (UID: "a0f7251c-fe97-4f8d-8121-82f7010170ef"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:09:44 crc kubenswrapper[4967]: I1008 00:09:44.646334 4967 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0f7251c-fe97-4f8d-8121-82f7010170ef-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 00:09:44 crc kubenswrapper[4967]: I1008 00:09:44.646367 4967 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0f7251c-fe97-4f8d-8121-82f7010170ef-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 00:09:44 crc kubenswrapper[4967]: I1008 00:09:44.646380 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t46zc\" (UniqueName: \"kubernetes.io/projected/a0f7251c-fe97-4f8d-8121-82f7010170ef-kube-api-access-t46zc\") on node \"crc\" DevicePath \"\"" Oct 08 00:09:44 crc kubenswrapper[4967]: I1008 00:09:44.838476 4967 generic.go:334] "Generic (PLEG): container finished" podID="a0f7251c-fe97-4f8d-8121-82f7010170ef" containerID="b9d92067efb26616bac31ab90d833e2ef6f3b8f02c1cc629588232e2a67ce396" exitCode=0 Oct 08 00:09:44 crc kubenswrapper[4967]: I1008 00:09:44.838548 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9sp7f" event={"ID":"a0f7251c-fe97-4f8d-8121-82f7010170ef","Type":"ContainerDied","Data":"b9d92067efb26616bac31ab90d833e2ef6f3b8f02c1cc629588232e2a67ce396"} Oct 08 00:09:44 crc kubenswrapper[4967]: I1008 00:09:44.838585 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9sp7f" event={"ID":"a0f7251c-fe97-4f8d-8121-82f7010170ef","Type":"ContainerDied","Data":"37e06fc47d07f2380a95cc6e276b8ffad84084d333e5067b79ba0544629f4642"} Oct 08 00:09:44 crc kubenswrapper[4967]: I1008 00:09:44.838614 4967 scope.go:117] "RemoveContainer" containerID="b9d92067efb26616bac31ab90d833e2ef6f3b8f02c1cc629588232e2a67ce396" Oct 08 00:09:44 crc kubenswrapper[4967]: I1008 00:09:44.838831 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9sp7f" Oct 08 00:09:44 crc kubenswrapper[4967]: I1008 00:09:44.865973 4967 scope.go:117] "RemoveContainer" containerID="6bbc46d8a8a19157a3c16a5fa630506dbaa7d6488a3cc69dc401687eca3d7b63" Oct 08 00:09:44 crc kubenswrapper[4967]: I1008 00:09:44.885143 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9sp7f"] Oct 08 00:09:44 crc kubenswrapper[4967]: I1008 00:09:44.887479 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-9sp7f"] Oct 08 00:09:44 crc kubenswrapper[4967]: I1008 00:09:44.908510 4967 scope.go:117] "RemoveContainer" containerID="62a0ec0ccaa70cfd198c426505e9901fdfc7ea35087b3cd9f7a306a7e954b9e1" Oct 08 00:09:44 crc kubenswrapper[4967]: I1008 00:09:44.928435 4967 scope.go:117] "RemoveContainer" containerID="b9d92067efb26616bac31ab90d833e2ef6f3b8f02c1cc629588232e2a67ce396" Oct 08 00:09:44 crc kubenswrapper[4967]: E1008 00:09:44.928996 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b9d92067efb26616bac31ab90d833e2ef6f3b8f02c1cc629588232e2a67ce396\": container with ID starting with b9d92067efb26616bac31ab90d833e2ef6f3b8f02c1cc629588232e2a67ce396 not found: ID does not exist" containerID="b9d92067efb26616bac31ab90d833e2ef6f3b8f02c1cc629588232e2a67ce396" Oct 08 00:09:44 crc kubenswrapper[4967]: I1008 00:09:44.929037 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9d92067efb26616bac31ab90d833e2ef6f3b8f02c1cc629588232e2a67ce396"} err="failed to get container status \"b9d92067efb26616bac31ab90d833e2ef6f3b8f02c1cc629588232e2a67ce396\": rpc error: code = NotFound desc = could not find container \"b9d92067efb26616bac31ab90d833e2ef6f3b8f02c1cc629588232e2a67ce396\": container with ID starting with b9d92067efb26616bac31ab90d833e2ef6f3b8f02c1cc629588232e2a67ce396 not found: ID does not exist" Oct 08 00:09:44 crc kubenswrapper[4967]: I1008 00:09:44.929089 4967 scope.go:117] "RemoveContainer" containerID="6bbc46d8a8a19157a3c16a5fa630506dbaa7d6488a3cc69dc401687eca3d7b63" Oct 08 00:09:44 crc kubenswrapper[4967]: E1008 00:09:44.930757 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6bbc46d8a8a19157a3c16a5fa630506dbaa7d6488a3cc69dc401687eca3d7b63\": container with ID starting with 6bbc46d8a8a19157a3c16a5fa630506dbaa7d6488a3cc69dc401687eca3d7b63 not found: ID does not exist" containerID="6bbc46d8a8a19157a3c16a5fa630506dbaa7d6488a3cc69dc401687eca3d7b63" Oct 08 00:09:44 crc kubenswrapper[4967]: I1008 00:09:44.930807 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6bbc46d8a8a19157a3c16a5fa630506dbaa7d6488a3cc69dc401687eca3d7b63"} err="failed to get container status \"6bbc46d8a8a19157a3c16a5fa630506dbaa7d6488a3cc69dc401687eca3d7b63\": rpc error: code = NotFound desc = could not find container \"6bbc46d8a8a19157a3c16a5fa630506dbaa7d6488a3cc69dc401687eca3d7b63\": container with ID starting with 6bbc46d8a8a19157a3c16a5fa630506dbaa7d6488a3cc69dc401687eca3d7b63 not found: ID does not exist" Oct 08 00:09:44 crc kubenswrapper[4967]: I1008 00:09:44.930841 4967 scope.go:117] "RemoveContainer" containerID="62a0ec0ccaa70cfd198c426505e9901fdfc7ea35087b3cd9f7a306a7e954b9e1" Oct 08 00:09:44 crc kubenswrapper[4967]: E1008 00:09:44.931228 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"62a0ec0ccaa70cfd198c426505e9901fdfc7ea35087b3cd9f7a306a7e954b9e1\": container with ID starting with 62a0ec0ccaa70cfd198c426505e9901fdfc7ea35087b3cd9f7a306a7e954b9e1 not found: ID does not exist" containerID="62a0ec0ccaa70cfd198c426505e9901fdfc7ea35087b3cd9f7a306a7e954b9e1" Oct 08 00:09:44 crc kubenswrapper[4967]: I1008 00:09:44.931254 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62a0ec0ccaa70cfd198c426505e9901fdfc7ea35087b3cd9f7a306a7e954b9e1"} err="failed to get container status \"62a0ec0ccaa70cfd198c426505e9901fdfc7ea35087b3cd9f7a306a7e954b9e1\": rpc error: code = NotFound desc = could not find container \"62a0ec0ccaa70cfd198c426505e9901fdfc7ea35087b3cd9f7a306a7e954b9e1\": container with ID starting with 62a0ec0ccaa70cfd198c426505e9901fdfc7ea35087b3cd9f7a306a7e954b9e1 not found: ID does not exist" Oct 08 00:09:46 crc kubenswrapper[4967]: I1008 00:09:46.124651 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0f7251c-fe97-4f8d-8121-82f7010170ef" path="/var/lib/kubelet/pods/a0f7251c-fe97-4f8d-8121-82f7010170ef/volumes" Oct 08 00:09:49 crc kubenswrapper[4967]: I1008 00:09:49.018351 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-9glfc" Oct 08 00:09:49 crc kubenswrapper[4967]: I1008 00:09:49.019301 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-9glfc" Oct 08 00:09:49 crc kubenswrapper[4967]: I1008 00:09:49.063640 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-9glfc" Oct 08 00:09:49 crc kubenswrapper[4967]: I1008 00:09:49.380303 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-x8d4p" Oct 08 00:09:49 crc kubenswrapper[4967]: I1008 00:09:49.459175 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-4hhcd" Oct 08 00:09:49 crc kubenswrapper[4967]: I1008 00:09:49.460115 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-4hhcd" Oct 08 00:09:49 crc kubenswrapper[4967]: I1008 00:09:49.495708 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-4hhcd" Oct 08 00:09:49 crc kubenswrapper[4967]: I1008 00:09:49.905212 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-9glfc" Oct 08 00:09:49 crc kubenswrapper[4967]: I1008 00:09:49.907060 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-4hhcd" Oct 08 00:09:51 crc kubenswrapper[4967]: I1008 00:09:51.459307 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-wcll6" Oct 08 00:09:51 crc kubenswrapper[4967]: I1008 00:09:51.552849 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4hhcd"] Oct 08 00:09:52 crc kubenswrapper[4967]: I1008 00:09:52.022703 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-v7scw" Oct 08 00:09:52 crc kubenswrapper[4967]: I1008 00:09:52.023039 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-v7scw" Oct 08 00:09:52 crc kubenswrapper[4967]: I1008 00:09:52.068103 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-v7scw" Oct 08 00:09:52 crc kubenswrapper[4967]: I1008 00:09:52.492529 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-f24p2" Oct 08 00:09:52 crc kubenswrapper[4967]: I1008 00:09:52.492876 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-f24p2" Oct 08 00:09:52 crc kubenswrapper[4967]: I1008 00:09:52.534043 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-f24p2" Oct 08 00:09:52 crc kubenswrapper[4967]: I1008 00:09:52.882976 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-4hhcd" podUID="8912b04d-fd70-497c-aab6-7490fc6ddf40" containerName="registry-server" containerID="cri-o://e7db1c6944d8b20adde1074eead4d289d796a9369b0a8ed3736e49af93449098" gracePeriod=2 Oct 08 00:09:52 crc kubenswrapper[4967]: I1008 00:09:52.924072 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-v7scw" Oct 08 00:09:52 crc kubenswrapper[4967]: I1008 00:09:52.926188 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-f24p2" Oct 08 00:09:53 crc kubenswrapper[4967]: I1008 00:09:53.756892 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wcll6"] Oct 08 00:09:53 crc kubenswrapper[4967]: I1008 00:09:53.759504 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-wcll6" podUID="9c827bbc-b90c-4168-918d-c2658da9b6ac" containerName="registry-server" containerID="cri-o://c0e93445fdd2997689d7ad3f47438f178238356a97d3da6986d3c0fd71645ea0" gracePeriod=2 Oct 08 00:09:53 crc kubenswrapper[4967]: I1008 00:09:53.891301 4967 generic.go:334] "Generic (PLEG): container finished" podID="8912b04d-fd70-497c-aab6-7490fc6ddf40" containerID="e7db1c6944d8b20adde1074eead4d289d796a9369b0a8ed3736e49af93449098" exitCode=0 Oct 08 00:09:53 crc kubenswrapper[4967]: I1008 00:09:53.891416 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4hhcd" event={"ID":"8912b04d-fd70-497c-aab6-7490fc6ddf40","Type":"ContainerDied","Data":"e7db1c6944d8b20adde1074eead4d289d796a9369b0a8ed3736e49af93449098"} Oct 08 00:09:53 crc kubenswrapper[4967]: I1008 00:09:53.891483 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4hhcd" event={"ID":"8912b04d-fd70-497c-aab6-7490fc6ddf40","Type":"ContainerDied","Data":"719b9ed2d135e5117f32eefac6f2b3426f69bddf531129a000cfc6e41aeb0848"} Oct 08 00:09:53 crc kubenswrapper[4967]: I1008 00:09:53.891509 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="719b9ed2d135e5117f32eefac6f2b3426f69bddf531129a000cfc6e41aeb0848" Oct 08 00:09:53 crc kubenswrapper[4967]: I1008 00:09:53.893629 4967 generic.go:334] "Generic (PLEG): container finished" podID="9c827bbc-b90c-4168-918d-c2658da9b6ac" containerID="c0e93445fdd2997689d7ad3f47438f178238356a97d3da6986d3c0fd71645ea0" exitCode=0 Oct 08 00:09:53 crc kubenswrapper[4967]: I1008 00:09:53.893709 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wcll6" event={"ID":"9c827bbc-b90c-4168-918d-c2658da9b6ac","Type":"ContainerDied","Data":"c0e93445fdd2997689d7ad3f47438f178238356a97d3da6986d3c0fd71645ea0"} Oct 08 00:09:53 crc kubenswrapper[4967]: I1008 00:09:53.929617 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4hhcd" Oct 08 00:09:54 crc kubenswrapper[4967]: I1008 00:09:54.064157 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8912b04d-fd70-497c-aab6-7490fc6ddf40-catalog-content\") pod \"8912b04d-fd70-497c-aab6-7490fc6ddf40\" (UID: \"8912b04d-fd70-497c-aab6-7490fc6ddf40\") " Oct 08 00:09:54 crc kubenswrapper[4967]: I1008 00:09:54.064215 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fwcx7\" (UniqueName: \"kubernetes.io/projected/8912b04d-fd70-497c-aab6-7490fc6ddf40-kube-api-access-fwcx7\") pod \"8912b04d-fd70-497c-aab6-7490fc6ddf40\" (UID: \"8912b04d-fd70-497c-aab6-7490fc6ddf40\") " Oct 08 00:09:54 crc kubenswrapper[4967]: I1008 00:09:54.064303 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8912b04d-fd70-497c-aab6-7490fc6ddf40-utilities\") pod \"8912b04d-fd70-497c-aab6-7490fc6ddf40\" (UID: \"8912b04d-fd70-497c-aab6-7490fc6ddf40\") " Oct 08 00:09:54 crc kubenswrapper[4967]: I1008 00:09:54.065108 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8912b04d-fd70-497c-aab6-7490fc6ddf40-utilities" (OuterVolumeSpecName: "utilities") pod "8912b04d-fd70-497c-aab6-7490fc6ddf40" (UID: "8912b04d-fd70-497c-aab6-7490fc6ddf40"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:09:54 crc kubenswrapper[4967]: I1008 00:09:54.091557 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8912b04d-fd70-497c-aab6-7490fc6ddf40-kube-api-access-fwcx7" (OuterVolumeSpecName: "kube-api-access-fwcx7") pod "8912b04d-fd70-497c-aab6-7490fc6ddf40" (UID: "8912b04d-fd70-497c-aab6-7490fc6ddf40"). InnerVolumeSpecName "kube-api-access-fwcx7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:09:54 crc kubenswrapper[4967]: I1008 00:09:54.131171 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8912b04d-fd70-497c-aab6-7490fc6ddf40-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8912b04d-fd70-497c-aab6-7490fc6ddf40" (UID: "8912b04d-fd70-497c-aab6-7490fc6ddf40"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:09:54 crc kubenswrapper[4967]: I1008 00:09:54.133384 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wcll6" Oct 08 00:09:54 crc kubenswrapper[4967]: I1008 00:09:54.167097 4967 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8912b04d-fd70-497c-aab6-7490fc6ddf40-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 00:09:54 crc kubenswrapper[4967]: I1008 00:09:54.167139 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fwcx7\" (UniqueName: \"kubernetes.io/projected/8912b04d-fd70-497c-aab6-7490fc6ddf40-kube-api-access-fwcx7\") on node \"crc\" DevicePath \"\"" Oct 08 00:09:54 crc kubenswrapper[4967]: I1008 00:09:54.167153 4967 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8912b04d-fd70-497c-aab6-7490fc6ddf40-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 00:09:54 crc kubenswrapper[4967]: I1008 00:09:54.267759 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c827bbc-b90c-4168-918d-c2658da9b6ac-catalog-content\") pod \"9c827bbc-b90c-4168-918d-c2658da9b6ac\" (UID: \"9c827bbc-b90c-4168-918d-c2658da9b6ac\") " Oct 08 00:09:54 crc kubenswrapper[4967]: I1008 00:09:54.267854 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zq8z8\" (UniqueName: \"kubernetes.io/projected/9c827bbc-b90c-4168-918d-c2658da9b6ac-kube-api-access-zq8z8\") pod \"9c827bbc-b90c-4168-918d-c2658da9b6ac\" (UID: \"9c827bbc-b90c-4168-918d-c2658da9b6ac\") " Oct 08 00:09:54 crc kubenswrapper[4967]: I1008 00:09:54.267897 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c827bbc-b90c-4168-918d-c2658da9b6ac-utilities\") pod \"9c827bbc-b90c-4168-918d-c2658da9b6ac\" (UID: \"9c827bbc-b90c-4168-918d-c2658da9b6ac\") " Oct 08 00:09:54 crc kubenswrapper[4967]: I1008 00:09:54.270366 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9c827bbc-b90c-4168-918d-c2658da9b6ac-utilities" (OuterVolumeSpecName: "utilities") pod "9c827bbc-b90c-4168-918d-c2658da9b6ac" (UID: "9c827bbc-b90c-4168-918d-c2658da9b6ac"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:09:54 crc kubenswrapper[4967]: I1008 00:09:54.280696 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c827bbc-b90c-4168-918d-c2658da9b6ac-kube-api-access-zq8z8" (OuterVolumeSpecName: "kube-api-access-zq8z8") pod "9c827bbc-b90c-4168-918d-c2658da9b6ac" (UID: "9c827bbc-b90c-4168-918d-c2658da9b6ac"). InnerVolumeSpecName "kube-api-access-zq8z8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:09:54 crc kubenswrapper[4967]: I1008 00:09:54.282630 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9c827bbc-b90c-4168-918d-c2658da9b6ac-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9c827bbc-b90c-4168-918d-c2658da9b6ac" (UID: "9c827bbc-b90c-4168-918d-c2658da9b6ac"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:09:54 crc kubenswrapper[4967]: I1008 00:09:54.379399 4967 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c827bbc-b90c-4168-918d-c2658da9b6ac-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 00:09:54 crc kubenswrapper[4967]: I1008 00:09:54.379447 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zq8z8\" (UniqueName: \"kubernetes.io/projected/9c827bbc-b90c-4168-918d-c2658da9b6ac-kube-api-access-zq8z8\") on node \"crc\" DevicePath \"\"" Oct 08 00:09:54 crc kubenswrapper[4967]: I1008 00:09:54.379459 4967 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c827bbc-b90c-4168-918d-c2658da9b6ac-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 00:09:54 crc kubenswrapper[4967]: I1008 00:09:54.900420 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wcll6" event={"ID":"9c827bbc-b90c-4168-918d-c2658da9b6ac","Type":"ContainerDied","Data":"8a6cbc39593c3932a978b35eb4d2adeeca4580b1fe025b83427766b4d5c28ae1"} Oct 08 00:09:54 crc kubenswrapper[4967]: I1008 00:09:54.900453 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wcll6" Oct 08 00:09:54 crc kubenswrapper[4967]: I1008 00:09:54.900486 4967 scope.go:117] "RemoveContainer" containerID="c0e93445fdd2997689d7ad3f47438f178238356a97d3da6986d3c0fd71645ea0" Oct 08 00:09:54 crc kubenswrapper[4967]: I1008 00:09:54.900453 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4hhcd" Oct 08 00:09:54 crc kubenswrapper[4967]: I1008 00:09:54.919697 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4hhcd"] Oct 08 00:09:54 crc kubenswrapper[4967]: I1008 00:09:54.921594 4967 scope.go:117] "RemoveContainer" containerID="a5db13822d13533dfbfddedd2e552ec1030b1ce84abd8a1c3250e83cabdbb5c1" Oct 08 00:09:54 crc kubenswrapper[4967]: I1008 00:09:54.923353 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-4hhcd"] Oct 08 00:09:54 crc kubenswrapper[4967]: I1008 00:09:54.932587 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wcll6"] Oct 08 00:09:54 crc kubenswrapper[4967]: I1008 00:09:54.938332 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-wcll6"] Oct 08 00:09:54 crc kubenswrapper[4967]: I1008 00:09:54.946715 4967 scope.go:117] "RemoveContainer" containerID="ed2206f1704b83fde92a9711faa9784858d66a94534908ef494a9a4864ba0a93" Oct 08 00:09:56 crc kubenswrapper[4967]: I1008 00:09:56.124520 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8912b04d-fd70-497c-aab6-7490fc6ddf40" path="/var/lib/kubelet/pods/8912b04d-fd70-497c-aab6-7490fc6ddf40/volumes" Oct 08 00:09:56 crc kubenswrapper[4967]: I1008 00:09:56.126355 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c827bbc-b90c-4168-918d-c2658da9b6ac" path="/var/lib/kubelet/pods/9c827bbc-b90c-4168-918d-c2658da9b6ac/volumes" Oct 08 00:09:56 crc kubenswrapper[4967]: I1008 00:09:56.154281 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-f24p2"] Oct 08 00:09:56 crc kubenswrapper[4967]: I1008 00:09:56.154544 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-f24p2" podUID="92a4f64f-e1cb-42a1-85b4-d3fc38e949a2" containerName="registry-server" containerID="cri-o://049fc4c8ddf143ec36fee0da8c5dd5c26592cca956aa969e73472747ab22e966" gracePeriod=2 Oct 08 00:09:56 crc kubenswrapper[4967]: I1008 00:09:56.483370 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f24p2" Oct 08 00:09:56 crc kubenswrapper[4967]: I1008 00:09:56.605535 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92a4f64f-e1cb-42a1-85b4-d3fc38e949a2-utilities\") pod \"92a4f64f-e1cb-42a1-85b4-d3fc38e949a2\" (UID: \"92a4f64f-e1cb-42a1-85b4-d3fc38e949a2\") " Oct 08 00:09:56 crc kubenswrapper[4967]: I1008 00:09:56.605706 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92a4f64f-e1cb-42a1-85b4-d3fc38e949a2-catalog-content\") pod \"92a4f64f-e1cb-42a1-85b4-d3fc38e949a2\" (UID: \"92a4f64f-e1cb-42a1-85b4-d3fc38e949a2\") " Oct 08 00:09:56 crc kubenswrapper[4967]: I1008 00:09:56.605745 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nf2kn\" (UniqueName: \"kubernetes.io/projected/92a4f64f-e1cb-42a1-85b4-d3fc38e949a2-kube-api-access-nf2kn\") pod \"92a4f64f-e1cb-42a1-85b4-d3fc38e949a2\" (UID: \"92a4f64f-e1cb-42a1-85b4-d3fc38e949a2\") " Oct 08 00:09:56 crc kubenswrapper[4967]: I1008 00:09:56.606519 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92a4f64f-e1cb-42a1-85b4-d3fc38e949a2-utilities" (OuterVolumeSpecName: "utilities") pod "92a4f64f-e1cb-42a1-85b4-d3fc38e949a2" (UID: "92a4f64f-e1cb-42a1-85b4-d3fc38e949a2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:09:56 crc kubenswrapper[4967]: I1008 00:09:56.612723 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92a4f64f-e1cb-42a1-85b4-d3fc38e949a2-kube-api-access-nf2kn" (OuterVolumeSpecName: "kube-api-access-nf2kn") pod "92a4f64f-e1cb-42a1-85b4-d3fc38e949a2" (UID: "92a4f64f-e1cb-42a1-85b4-d3fc38e949a2"). InnerVolumeSpecName "kube-api-access-nf2kn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:09:56 crc kubenswrapper[4967]: I1008 00:09:56.690035 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92a4f64f-e1cb-42a1-85b4-d3fc38e949a2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "92a4f64f-e1cb-42a1-85b4-d3fc38e949a2" (UID: "92a4f64f-e1cb-42a1-85b4-d3fc38e949a2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:09:56 crc kubenswrapper[4967]: I1008 00:09:56.707525 4967 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92a4f64f-e1cb-42a1-85b4-d3fc38e949a2-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 00:09:56 crc kubenswrapper[4967]: I1008 00:09:56.707571 4967 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92a4f64f-e1cb-42a1-85b4-d3fc38e949a2-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 00:09:56 crc kubenswrapper[4967]: I1008 00:09:56.707589 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nf2kn\" (UniqueName: \"kubernetes.io/projected/92a4f64f-e1cb-42a1-85b4-d3fc38e949a2-kube-api-access-nf2kn\") on node \"crc\" DevicePath \"\"" Oct 08 00:09:56 crc kubenswrapper[4967]: I1008 00:09:56.914782 4967 generic.go:334] "Generic (PLEG): container finished" podID="92a4f64f-e1cb-42a1-85b4-d3fc38e949a2" containerID="049fc4c8ddf143ec36fee0da8c5dd5c26592cca956aa969e73472747ab22e966" exitCode=0 Oct 08 00:09:56 crc kubenswrapper[4967]: I1008 00:09:56.914832 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f24p2" event={"ID":"92a4f64f-e1cb-42a1-85b4-d3fc38e949a2","Type":"ContainerDied","Data":"049fc4c8ddf143ec36fee0da8c5dd5c26592cca956aa969e73472747ab22e966"} Oct 08 00:09:56 crc kubenswrapper[4967]: I1008 00:09:56.914839 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-f24p2" Oct 08 00:09:56 crc kubenswrapper[4967]: I1008 00:09:56.914864 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-f24p2" event={"ID":"92a4f64f-e1cb-42a1-85b4-d3fc38e949a2","Type":"ContainerDied","Data":"5765d732093b5e0be047c8db149b7dbb5008bcb5ed58e81a6284b667cbb7121b"} Oct 08 00:09:56 crc kubenswrapper[4967]: I1008 00:09:56.914889 4967 scope.go:117] "RemoveContainer" containerID="049fc4c8ddf143ec36fee0da8c5dd5c26592cca956aa969e73472747ab22e966" Oct 08 00:09:56 crc kubenswrapper[4967]: I1008 00:09:56.931448 4967 scope.go:117] "RemoveContainer" containerID="d013ec6e24839f4a89cbc8aea45a44ca43c39887969b94e9b9e0e452fceace04" Oct 08 00:09:56 crc kubenswrapper[4967]: I1008 00:09:56.943864 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-f24p2"] Oct 08 00:09:56 crc kubenswrapper[4967]: I1008 00:09:56.947359 4967 scope.go:117] "RemoveContainer" containerID="ec33944e8ccb6004145beaaa46539b7be44284a9a6c3afd0c283efdc6ab8c569" Oct 08 00:09:56 crc kubenswrapper[4967]: I1008 00:09:56.949808 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-f24p2"] Oct 08 00:09:56 crc kubenswrapper[4967]: I1008 00:09:56.964086 4967 scope.go:117] "RemoveContainer" containerID="049fc4c8ddf143ec36fee0da8c5dd5c26592cca956aa969e73472747ab22e966" Oct 08 00:09:56 crc kubenswrapper[4967]: E1008 00:09:56.964729 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"049fc4c8ddf143ec36fee0da8c5dd5c26592cca956aa969e73472747ab22e966\": container with ID starting with 049fc4c8ddf143ec36fee0da8c5dd5c26592cca956aa969e73472747ab22e966 not found: ID does not exist" containerID="049fc4c8ddf143ec36fee0da8c5dd5c26592cca956aa969e73472747ab22e966" Oct 08 00:09:56 crc kubenswrapper[4967]: I1008 00:09:56.964780 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"049fc4c8ddf143ec36fee0da8c5dd5c26592cca956aa969e73472747ab22e966"} err="failed to get container status \"049fc4c8ddf143ec36fee0da8c5dd5c26592cca956aa969e73472747ab22e966\": rpc error: code = NotFound desc = could not find container \"049fc4c8ddf143ec36fee0da8c5dd5c26592cca956aa969e73472747ab22e966\": container with ID starting with 049fc4c8ddf143ec36fee0da8c5dd5c26592cca956aa969e73472747ab22e966 not found: ID does not exist" Oct 08 00:09:56 crc kubenswrapper[4967]: I1008 00:09:56.964815 4967 scope.go:117] "RemoveContainer" containerID="d013ec6e24839f4a89cbc8aea45a44ca43c39887969b94e9b9e0e452fceace04" Oct 08 00:09:56 crc kubenswrapper[4967]: E1008 00:09:56.965203 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d013ec6e24839f4a89cbc8aea45a44ca43c39887969b94e9b9e0e452fceace04\": container with ID starting with d013ec6e24839f4a89cbc8aea45a44ca43c39887969b94e9b9e0e452fceace04 not found: ID does not exist" containerID="d013ec6e24839f4a89cbc8aea45a44ca43c39887969b94e9b9e0e452fceace04" Oct 08 00:09:56 crc kubenswrapper[4967]: I1008 00:09:56.965250 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d013ec6e24839f4a89cbc8aea45a44ca43c39887969b94e9b9e0e452fceace04"} err="failed to get container status \"d013ec6e24839f4a89cbc8aea45a44ca43c39887969b94e9b9e0e452fceace04\": rpc error: code = NotFound desc = could not find container \"d013ec6e24839f4a89cbc8aea45a44ca43c39887969b94e9b9e0e452fceace04\": container with ID starting with d013ec6e24839f4a89cbc8aea45a44ca43c39887969b94e9b9e0e452fceace04 not found: ID does not exist" Oct 08 00:09:56 crc kubenswrapper[4967]: I1008 00:09:56.965288 4967 scope.go:117] "RemoveContainer" containerID="ec33944e8ccb6004145beaaa46539b7be44284a9a6c3afd0c283efdc6ab8c569" Oct 08 00:09:56 crc kubenswrapper[4967]: E1008 00:09:56.965609 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec33944e8ccb6004145beaaa46539b7be44284a9a6c3afd0c283efdc6ab8c569\": container with ID starting with ec33944e8ccb6004145beaaa46539b7be44284a9a6c3afd0c283efdc6ab8c569 not found: ID does not exist" containerID="ec33944e8ccb6004145beaaa46539b7be44284a9a6c3afd0c283efdc6ab8c569" Oct 08 00:09:56 crc kubenswrapper[4967]: I1008 00:09:56.965644 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec33944e8ccb6004145beaaa46539b7be44284a9a6c3afd0c283efdc6ab8c569"} err="failed to get container status \"ec33944e8ccb6004145beaaa46539b7be44284a9a6c3afd0c283efdc6ab8c569\": rpc error: code = NotFound desc = could not find container \"ec33944e8ccb6004145beaaa46539b7be44284a9a6c3afd0c283efdc6ab8c569\": container with ID starting with ec33944e8ccb6004145beaaa46539b7be44284a9a6c3afd0c283efdc6ab8c569 not found: ID does not exist" Oct 08 00:09:58 crc kubenswrapper[4967]: I1008 00:09:58.128877 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92a4f64f-e1cb-42a1-85b4-d3fc38e949a2" path="/var/lib/kubelet/pods/92a4f64f-e1cb-42a1-85b4-d3fc38e949a2/volumes" Oct 08 00:10:01 crc kubenswrapper[4967]: I1008 00:10:01.378788 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 00:10:02 crc kubenswrapper[4967]: I1008 00:10:02.080412 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-tmj7c"] Oct 08 00:10:06 crc kubenswrapper[4967]: I1008 00:10:06.920398 4967 patch_prober.go:28] interesting pod/machine-config-daemon-xvdtp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 00:10:06 crc kubenswrapper[4967]: I1008 00:10:06.921137 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" podUID="9461162a-c27d-465e-9ca7-2d3936877cd8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 00:10:06 crc kubenswrapper[4967]: I1008 00:10:06.921207 4967 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" Oct 08 00:10:06 crc kubenswrapper[4967]: I1008 00:10:06.922109 4967 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"28760af04ef321d0f91b2bce3ad074d44b63dabf676cc37f3be6aa8726341a2f"} pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 00:10:06 crc kubenswrapper[4967]: I1008 00:10:06.922195 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" podUID="9461162a-c27d-465e-9ca7-2d3936877cd8" containerName="machine-config-daemon" containerID="cri-o://28760af04ef321d0f91b2bce3ad074d44b63dabf676cc37f3be6aa8726341a2f" gracePeriod=600 Oct 08 00:10:07 crc kubenswrapper[4967]: I1008 00:10:07.984424 4967 generic.go:334] "Generic (PLEG): container finished" podID="9461162a-c27d-465e-9ca7-2d3936877cd8" containerID="28760af04ef321d0f91b2bce3ad074d44b63dabf676cc37f3be6aa8726341a2f" exitCode=0 Oct 08 00:10:07 crc kubenswrapper[4967]: I1008 00:10:07.984580 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" event={"ID":"9461162a-c27d-465e-9ca7-2d3936877cd8","Type":"ContainerDied","Data":"28760af04ef321d0f91b2bce3ad074d44b63dabf676cc37f3be6aa8726341a2f"} Oct 08 00:10:07 crc kubenswrapper[4967]: I1008 00:10:07.984959 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" event={"ID":"9461162a-c27d-465e-9ca7-2d3936877cd8","Type":"ContainerStarted","Data":"d566016b7da391d83d191f50824299b48d2b961375cdeea4e96f5ac21ff9991f"} Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.113880 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-tmj7c" podUID="d7ec4b5e-d6b4-4681-a231-c7811661c6cb" containerName="oauth-openshift" containerID="cri-o://30495094bf636d17573cf31907c204329ec27e29d279878d6eeecc6249b8a486" gracePeriod=15 Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.539006 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-tmj7c" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.579768 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-76766fc778-hl47l"] Oct 08 00:10:27 crc kubenswrapper[4967]: E1008 00:10:27.580482 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea71a082-3a10-42af-b39c-69ec1c0eb242" containerName="pruner" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.580767 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea71a082-3a10-42af-b39c-69ec1c0eb242" containerName="pruner" Oct 08 00:10:27 crc kubenswrapper[4967]: E1008 00:10:27.580982 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8912b04d-fd70-497c-aab6-7490fc6ddf40" containerName="extract-utilities" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.581159 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="8912b04d-fd70-497c-aab6-7490fc6ddf40" containerName="extract-utilities" Oct 08 00:10:27 crc kubenswrapper[4967]: E1008 00:10:27.581196 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8912b04d-fd70-497c-aab6-7490fc6ddf40" containerName="extract-content" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.581214 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="8912b04d-fd70-497c-aab6-7490fc6ddf40" containerName="extract-content" Oct 08 00:10:27 crc kubenswrapper[4967]: E1008 00:10:27.581242 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c827bbc-b90c-4168-918d-c2658da9b6ac" containerName="extract-utilities" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.581258 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c827bbc-b90c-4168-918d-c2658da9b6ac" containerName="extract-utilities" Oct 08 00:10:27 crc kubenswrapper[4967]: E1008 00:10:27.581283 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb260fdf-16d6-4771-b7ae-38a669ff2393" containerName="collect-profiles" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.581298 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb260fdf-16d6-4771-b7ae-38a669ff2393" containerName="collect-profiles" Oct 08 00:10:27 crc kubenswrapper[4967]: E1008 00:10:27.581319 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92a4f64f-e1cb-42a1-85b4-d3fc38e949a2" containerName="extract-utilities" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.581337 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="92a4f64f-e1cb-42a1-85b4-d3fc38e949a2" containerName="extract-utilities" Oct 08 00:10:27 crc kubenswrapper[4967]: E1008 00:10:27.581357 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c827bbc-b90c-4168-918d-c2658da9b6ac" containerName="extract-content" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.581372 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c827bbc-b90c-4168-918d-c2658da9b6ac" containerName="extract-content" Oct 08 00:10:27 crc kubenswrapper[4967]: E1008 00:10:27.581395 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0f7251c-fe97-4f8d-8121-82f7010170ef" containerName="registry-server" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.581410 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0f7251c-fe97-4f8d-8121-82f7010170ef" containerName="registry-server" Oct 08 00:10:27 crc kubenswrapper[4967]: E1008 00:10:27.581433 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92a4f64f-e1cb-42a1-85b4-d3fc38e949a2" containerName="registry-server" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.581448 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="92a4f64f-e1cb-42a1-85b4-d3fc38e949a2" containerName="registry-server" Oct 08 00:10:27 crc kubenswrapper[4967]: E1008 00:10:27.581474 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0f7251c-fe97-4f8d-8121-82f7010170ef" containerName="extract-content" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.581489 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0f7251c-fe97-4f8d-8121-82f7010170ef" containerName="extract-content" Oct 08 00:10:27 crc kubenswrapper[4967]: E1008 00:10:27.581510 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0f7251c-fe97-4f8d-8121-82f7010170ef" containerName="extract-utilities" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.581525 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0f7251c-fe97-4f8d-8121-82f7010170ef" containerName="extract-utilities" Oct 08 00:10:27 crc kubenswrapper[4967]: E1008 00:10:27.581549 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c827bbc-b90c-4168-918d-c2658da9b6ac" containerName="registry-server" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.581567 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c827bbc-b90c-4168-918d-c2658da9b6ac" containerName="registry-server" Oct 08 00:10:27 crc kubenswrapper[4967]: E1008 00:10:27.581605 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92a4f64f-e1cb-42a1-85b4-d3fc38e949a2" containerName="extract-content" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.581624 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="92a4f64f-e1cb-42a1-85b4-d3fc38e949a2" containerName="extract-content" Oct 08 00:10:27 crc kubenswrapper[4967]: E1008 00:10:27.581641 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7ec4b5e-d6b4-4681-a231-c7811661c6cb" containerName="oauth-openshift" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.581657 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7ec4b5e-d6b4-4681-a231-c7811661c6cb" containerName="oauth-openshift" Oct 08 00:10:27 crc kubenswrapper[4967]: E1008 00:10:27.581719 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84cd9a7a-bb01-43de-9055-deccbc8529b2" containerName="image-pruner" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.581738 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="84cd9a7a-bb01-43de-9055-deccbc8529b2" containerName="image-pruner" Oct 08 00:10:27 crc kubenswrapper[4967]: E1008 00:10:27.581759 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8912b04d-fd70-497c-aab6-7490fc6ddf40" containerName="registry-server" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.581774 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="8912b04d-fd70-497c-aab6-7490fc6ddf40" containerName="registry-server" Oct 08 00:10:27 crc kubenswrapper[4967]: E1008 00:10:27.581799 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85a3da15-9895-4e7f-81f4-6be5e34061de" containerName="pruner" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.581814 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="85a3da15-9895-4e7f-81f4-6be5e34061de" containerName="pruner" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.582034 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="85a3da15-9895-4e7f-81f4-6be5e34061de" containerName="pruner" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.582071 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0f7251c-fe97-4f8d-8121-82f7010170ef" containerName="registry-server" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.582097 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="92a4f64f-e1cb-42a1-85b4-d3fc38e949a2" containerName="registry-server" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.582119 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c827bbc-b90c-4168-918d-c2658da9b6ac" containerName="registry-server" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.582137 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7ec4b5e-d6b4-4681-a231-c7811661c6cb" containerName="oauth-openshift" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.582155 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb260fdf-16d6-4771-b7ae-38a669ff2393" containerName="collect-profiles" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.582178 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="84cd9a7a-bb01-43de-9055-deccbc8529b2" containerName="image-pruner" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.582196 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea71a082-3a10-42af-b39c-69ec1c0eb242" containerName="pruner" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.582222 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="8912b04d-fd70-497c-aab6-7490fc6ddf40" containerName="registry-server" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.582915 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-76766fc778-hl47l" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.593279 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-76766fc778-hl47l"] Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.726205 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-v4-0-config-system-session\") pod \"d7ec4b5e-d6b4-4681-a231-c7811661c6cb\" (UID: \"d7ec4b5e-d6b4-4681-a231-c7811661c6cb\") " Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.726273 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-v4-0-config-user-template-error\") pod \"d7ec4b5e-d6b4-4681-a231-c7811661c6cb\" (UID: \"d7ec4b5e-d6b4-4681-a231-c7811661c6cb\") " Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.726335 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-v4-0-config-system-ocp-branding-template\") pod \"d7ec4b5e-d6b4-4681-a231-c7811661c6cb\" (UID: \"d7ec4b5e-d6b4-4681-a231-c7811661c6cb\") " Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.726357 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-v4-0-config-system-serving-cert\") pod \"d7ec4b5e-d6b4-4681-a231-c7811661c6cb\" (UID: \"d7ec4b5e-d6b4-4681-a231-c7811661c6cb\") " Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.726405 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-audit-policies\") pod \"d7ec4b5e-d6b4-4681-a231-c7811661c6cb\" (UID: \"d7ec4b5e-d6b4-4681-a231-c7811661c6cb\") " Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.726433 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-v4-0-config-user-template-login\") pod \"d7ec4b5e-d6b4-4681-a231-c7811661c6cb\" (UID: \"d7ec4b5e-d6b4-4681-a231-c7811661c6cb\") " Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.726472 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-v4-0-config-user-template-provider-selection\") pod \"d7ec4b5e-d6b4-4681-a231-c7811661c6cb\" (UID: \"d7ec4b5e-d6b4-4681-a231-c7811661c6cb\") " Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.726495 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-v4-0-config-system-cliconfig\") pod \"d7ec4b5e-d6b4-4681-a231-c7811661c6cb\" (UID: \"d7ec4b5e-d6b4-4681-a231-c7811661c6cb\") " Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.726515 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-v4-0-config-system-trusted-ca-bundle\") pod \"d7ec4b5e-d6b4-4681-a231-c7811661c6cb\" (UID: \"d7ec4b5e-d6b4-4681-a231-c7811661c6cb\") " Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.726570 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-v4-0-config-system-service-ca\") pod \"d7ec4b5e-d6b4-4681-a231-c7811661c6cb\" (UID: \"d7ec4b5e-d6b4-4681-a231-c7811661c6cb\") " Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.726618 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-v4-0-config-user-idp-0-file-data\") pod \"d7ec4b5e-d6b4-4681-a231-c7811661c6cb\" (UID: \"d7ec4b5e-d6b4-4681-a231-c7811661c6cb\") " Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.726650 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-audit-dir\") pod \"d7ec4b5e-d6b4-4681-a231-c7811661c6cb\" (UID: \"d7ec4b5e-d6b4-4681-a231-c7811661c6cb\") " Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.726694 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mhcpg\" (UniqueName: \"kubernetes.io/projected/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-kube-api-access-mhcpg\") pod \"d7ec4b5e-d6b4-4681-a231-c7811661c6cb\" (UID: \"d7ec4b5e-d6b4-4681-a231-c7811661c6cb\") " Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.726719 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-v4-0-config-system-router-certs\") pod \"d7ec4b5e-d6b4-4681-a231-c7811661c6cb\" (UID: \"d7ec4b5e-d6b4-4681-a231-c7811661c6cb\") " Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.726862 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/ebb960b7-8b92-4538-a05f-d0a4460c0fc7-v4-0-config-system-router-certs\") pod \"oauth-openshift-76766fc778-hl47l\" (UID: \"ebb960b7-8b92-4538-a05f-d0a4460c0fc7\") " pod="openshift-authentication/oauth-openshift-76766fc778-hl47l" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.726896 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6d2w6\" (UniqueName: \"kubernetes.io/projected/ebb960b7-8b92-4538-a05f-d0a4460c0fc7-kube-api-access-6d2w6\") pod \"oauth-openshift-76766fc778-hl47l\" (UID: \"ebb960b7-8b92-4538-a05f-d0a4460c0fc7\") " pod="openshift-authentication/oauth-openshift-76766fc778-hl47l" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.726936 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ebb960b7-8b92-4538-a05f-d0a4460c0fc7-audit-policies\") pod \"oauth-openshift-76766fc778-hl47l\" (UID: \"ebb960b7-8b92-4538-a05f-d0a4460c0fc7\") " pod="openshift-authentication/oauth-openshift-76766fc778-hl47l" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.726952 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/ebb960b7-8b92-4538-a05f-d0a4460c0fc7-v4-0-config-system-serving-cert\") pod \"oauth-openshift-76766fc778-hl47l\" (UID: \"ebb960b7-8b92-4538-a05f-d0a4460c0fc7\") " pod="openshift-authentication/oauth-openshift-76766fc778-hl47l" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.726979 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/ebb960b7-8b92-4538-a05f-d0a4460c0fc7-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-76766fc778-hl47l\" (UID: \"ebb960b7-8b92-4538-a05f-d0a4460c0fc7\") " pod="openshift-authentication/oauth-openshift-76766fc778-hl47l" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.727030 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/ebb960b7-8b92-4538-a05f-d0a4460c0fc7-v4-0-config-system-service-ca\") pod \"oauth-openshift-76766fc778-hl47l\" (UID: \"ebb960b7-8b92-4538-a05f-d0a4460c0fc7\") " pod="openshift-authentication/oauth-openshift-76766fc778-hl47l" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.727053 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/ebb960b7-8b92-4538-a05f-d0a4460c0fc7-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-76766fc778-hl47l\" (UID: \"ebb960b7-8b92-4538-a05f-d0a4460c0fc7\") " pod="openshift-authentication/oauth-openshift-76766fc778-hl47l" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.727117 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/ebb960b7-8b92-4538-a05f-d0a4460c0fc7-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-76766fc778-hl47l\" (UID: \"ebb960b7-8b92-4538-a05f-d0a4460c0fc7\") " pod="openshift-authentication/oauth-openshift-76766fc778-hl47l" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.727180 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/ebb960b7-8b92-4538-a05f-d0a4460c0fc7-v4-0-config-system-session\") pod \"oauth-openshift-76766fc778-hl47l\" (UID: \"ebb960b7-8b92-4538-a05f-d0a4460c0fc7\") " pod="openshift-authentication/oauth-openshift-76766fc778-hl47l" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.727210 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/ebb960b7-8b92-4538-a05f-d0a4460c0fc7-v4-0-config-system-cliconfig\") pod \"oauth-openshift-76766fc778-hl47l\" (UID: \"ebb960b7-8b92-4538-a05f-d0a4460c0fc7\") " pod="openshift-authentication/oauth-openshift-76766fc778-hl47l" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.727255 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ebb960b7-8b92-4538-a05f-d0a4460c0fc7-audit-dir\") pod \"oauth-openshift-76766fc778-hl47l\" (UID: \"ebb960b7-8b92-4538-a05f-d0a4460c0fc7\") " pod="openshift-authentication/oauth-openshift-76766fc778-hl47l" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.727275 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/ebb960b7-8b92-4538-a05f-d0a4460c0fc7-v4-0-config-user-template-error\") pod \"oauth-openshift-76766fc778-hl47l\" (UID: \"ebb960b7-8b92-4538-a05f-d0a4460c0fc7\") " pod="openshift-authentication/oauth-openshift-76766fc778-hl47l" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.727303 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/ebb960b7-8b92-4538-a05f-d0a4460c0fc7-v4-0-config-user-template-login\") pod \"oauth-openshift-76766fc778-hl47l\" (UID: \"ebb960b7-8b92-4538-a05f-d0a4460c0fc7\") " pod="openshift-authentication/oauth-openshift-76766fc778-hl47l" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.727342 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ebb960b7-8b92-4538-a05f-d0a4460c0fc7-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-76766fc778-hl47l\" (UID: \"ebb960b7-8b92-4538-a05f-d0a4460c0fc7\") " pod="openshift-authentication/oauth-openshift-76766fc778-hl47l" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.729374 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "d7ec4b5e-d6b4-4681-a231-c7811661c6cb" (UID: "d7ec4b5e-d6b4-4681-a231-c7811661c6cb"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.729582 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "d7ec4b5e-d6b4-4681-a231-c7811661c6cb" (UID: "d7ec4b5e-d6b4-4681-a231-c7811661c6cb"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.730096 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "d7ec4b5e-d6b4-4681-a231-c7811661c6cb" (UID: "d7ec4b5e-d6b4-4681-a231-c7811661c6cb"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.730447 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "d7ec4b5e-d6b4-4681-a231-c7811661c6cb" (UID: "d7ec4b5e-d6b4-4681-a231-c7811661c6cb"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.733872 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "d7ec4b5e-d6b4-4681-a231-c7811661c6cb" (UID: "d7ec4b5e-d6b4-4681-a231-c7811661c6cb"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.734315 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "d7ec4b5e-d6b4-4681-a231-c7811661c6cb" (UID: "d7ec4b5e-d6b4-4681-a231-c7811661c6cb"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.736488 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "d7ec4b5e-d6b4-4681-a231-c7811661c6cb" (UID: "d7ec4b5e-d6b4-4681-a231-c7811661c6cb"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.738231 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-kube-api-access-mhcpg" (OuterVolumeSpecName: "kube-api-access-mhcpg") pod "d7ec4b5e-d6b4-4681-a231-c7811661c6cb" (UID: "d7ec4b5e-d6b4-4681-a231-c7811661c6cb"). InnerVolumeSpecName "kube-api-access-mhcpg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.738841 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "d7ec4b5e-d6b4-4681-a231-c7811661c6cb" (UID: "d7ec4b5e-d6b4-4681-a231-c7811661c6cb"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.739591 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "d7ec4b5e-d6b4-4681-a231-c7811661c6cb" (UID: "d7ec4b5e-d6b4-4681-a231-c7811661c6cb"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.744626 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "d7ec4b5e-d6b4-4681-a231-c7811661c6cb" (UID: "d7ec4b5e-d6b4-4681-a231-c7811661c6cb"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.749729 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "d7ec4b5e-d6b4-4681-a231-c7811661c6cb" (UID: "d7ec4b5e-d6b4-4681-a231-c7811661c6cb"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.751168 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "d7ec4b5e-d6b4-4681-a231-c7811661c6cb" (UID: "d7ec4b5e-d6b4-4681-a231-c7811661c6cb"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.751325 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "d7ec4b5e-d6b4-4681-a231-c7811661c6cb" (UID: "d7ec4b5e-d6b4-4681-a231-c7811661c6cb"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.829008 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/ebb960b7-8b92-4538-a05f-d0a4460c0fc7-v4-0-config-system-service-ca\") pod \"oauth-openshift-76766fc778-hl47l\" (UID: \"ebb960b7-8b92-4538-a05f-d0a4460c0fc7\") " pod="openshift-authentication/oauth-openshift-76766fc778-hl47l" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.829107 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/ebb960b7-8b92-4538-a05f-d0a4460c0fc7-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-76766fc778-hl47l\" (UID: \"ebb960b7-8b92-4538-a05f-d0a4460c0fc7\") " pod="openshift-authentication/oauth-openshift-76766fc778-hl47l" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.829158 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/ebb960b7-8b92-4538-a05f-d0a4460c0fc7-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-76766fc778-hl47l\" (UID: \"ebb960b7-8b92-4538-a05f-d0a4460c0fc7\") " pod="openshift-authentication/oauth-openshift-76766fc778-hl47l" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.829202 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/ebb960b7-8b92-4538-a05f-d0a4460c0fc7-v4-0-config-system-session\") pod \"oauth-openshift-76766fc778-hl47l\" (UID: \"ebb960b7-8b92-4538-a05f-d0a4460c0fc7\") " pod="openshift-authentication/oauth-openshift-76766fc778-hl47l" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.829233 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/ebb960b7-8b92-4538-a05f-d0a4460c0fc7-v4-0-config-system-cliconfig\") pod \"oauth-openshift-76766fc778-hl47l\" (UID: \"ebb960b7-8b92-4538-a05f-d0a4460c0fc7\") " pod="openshift-authentication/oauth-openshift-76766fc778-hl47l" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.829273 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ebb960b7-8b92-4538-a05f-d0a4460c0fc7-audit-dir\") pod \"oauth-openshift-76766fc778-hl47l\" (UID: \"ebb960b7-8b92-4538-a05f-d0a4460c0fc7\") " pod="openshift-authentication/oauth-openshift-76766fc778-hl47l" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.829327 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/ebb960b7-8b92-4538-a05f-d0a4460c0fc7-v4-0-config-user-template-error\") pod \"oauth-openshift-76766fc778-hl47l\" (UID: \"ebb960b7-8b92-4538-a05f-d0a4460c0fc7\") " pod="openshift-authentication/oauth-openshift-76766fc778-hl47l" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.829391 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/ebb960b7-8b92-4538-a05f-d0a4460c0fc7-v4-0-config-user-template-login\") pod \"oauth-openshift-76766fc778-hl47l\" (UID: \"ebb960b7-8b92-4538-a05f-d0a4460c0fc7\") " pod="openshift-authentication/oauth-openshift-76766fc778-hl47l" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.829425 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ebb960b7-8b92-4538-a05f-d0a4460c0fc7-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-76766fc778-hl47l\" (UID: \"ebb960b7-8b92-4538-a05f-d0a4460c0fc7\") " pod="openshift-authentication/oauth-openshift-76766fc778-hl47l" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.829468 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/ebb960b7-8b92-4538-a05f-d0a4460c0fc7-v4-0-config-system-router-certs\") pod \"oauth-openshift-76766fc778-hl47l\" (UID: \"ebb960b7-8b92-4538-a05f-d0a4460c0fc7\") " pod="openshift-authentication/oauth-openshift-76766fc778-hl47l" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.829522 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6d2w6\" (UniqueName: \"kubernetes.io/projected/ebb960b7-8b92-4538-a05f-d0a4460c0fc7-kube-api-access-6d2w6\") pod \"oauth-openshift-76766fc778-hl47l\" (UID: \"ebb960b7-8b92-4538-a05f-d0a4460c0fc7\") " pod="openshift-authentication/oauth-openshift-76766fc778-hl47l" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.829558 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ebb960b7-8b92-4538-a05f-d0a4460c0fc7-audit-policies\") pod \"oauth-openshift-76766fc778-hl47l\" (UID: \"ebb960b7-8b92-4538-a05f-d0a4460c0fc7\") " pod="openshift-authentication/oauth-openshift-76766fc778-hl47l" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.829593 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/ebb960b7-8b92-4538-a05f-d0a4460c0fc7-v4-0-config-system-serving-cert\") pod \"oauth-openshift-76766fc778-hl47l\" (UID: \"ebb960b7-8b92-4538-a05f-d0a4460c0fc7\") " pod="openshift-authentication/oauth-openshift-76766fc778-hl47l" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.829638 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/ebb960b7-8b92-4538-a05f-d0a4460c0fc7-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-76766fc778-hl47l\" (UID: \"ebb960b7-8b92-4538-a05f-d0a4460c0fc7\") " pod="openshift-authentication/oauth-openshift-76766fc778-hl47l" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.829760 4967 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.829784 4967 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.829809 4967 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.829830 4967 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.829853 4967 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-audit-dir\") on node \"crc\" DevicePath \"\"" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.829873 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mhcpg\" (UniqueName: \"kubernetes.io/projected/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-kube-api-access-mhcpg\") on node \"crc\" DevicePath \"\"" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.829894 4967 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.829913 4967 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.829933 4967 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.830042 4967 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.830075 4967 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.830140 4967 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.830164 4967 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.830185 4967 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/d7ec4b5e-d6b4-4681-a231-c7811661c6cb-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.831602 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/ebb960b7-8b92-4538-a05f-d0a4460c0fc7-v4-0-config-system-service-ca\") pod \"oauth-openshift-76766fc778-hl47l\" (UID: \"ebb960b7-8b92-4538-a05f-d0a4460c0fc7\") " pod="openshift-authentication/oauth-openshift-76766fc778-hl47l" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.833175 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ebb960b7-8b92-4538-a05f-d0a4460c0fc7-audit-policies\") pod \"oauth-openshift-76766fc778-hl47l\" (UID: \"ebb960b7-8b92-4538-a05f-d0a4460c0fc7\") " pod="openshift-authentication/oauth-openshift-76766fc778-hl47l" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.833318 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ebb960b7-8b92-4538-a05f-d0a4460c0fc7-audit-dir\") pod \"oauth-openshift-76766fc778-hl47l\" (UID: \"ebb960b7-8b92-4538-a05f-d0a4460c0fc7\") " pod="openshift-authentication/oauth-openshift-76766fc778-hl47l" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.834035 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ebb960b7-8b92-4538-a05f-d0a4460c0fc7-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-76766fc778-hl47l\" (UID: \"ebb960b7-8b92-4538-a05f-d0a4460c0fc7\") " pod="openshift-authentication/oauth-openshift-76766fc778-hl47l" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.836181 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/ebb960b7-8b92-4538-a05f-d0a4460c0fc7-v4-0-config-system-cliconfig\") pod \"oauth-openshift-76766fc778-hl47l\" (UID: \"ebb960b7-8b92-4538-a05f-d0a4460c0fc7\") " pod="openshift-authentication/oauth-openshift-76766fc778-hl47l" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.837000 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/ebb960b7-8b92-4538-a05f-d0a4460c0fc7-v4-0-config-user-template-login\") pod \"oauth-openshift-76766fc778-hl47l\" (UID: \"ebb960b7-8b92-4538-a05f-d0a4460c0fc7\") " pod="openshift-authentication/oauth-openshift-76766fc778-hl47l" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.837216 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/ebb960b7-8b92-4538-a05f-d0a4460c0fc7-v4-0-config-system-session\") pod \"oauth-openshift-76766fc778-hl47l\" (UID: \"ebb960b7-8b92-4538-a05f-d0a4460c0fc7\") " pod="openshift-authentication/oauth-openshift-76766fc778-hl47l" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.837621 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/ebb960b7-8b92-4538-a05f-d0a4460c0fc7-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-76766fc778-hl47l\" (UID: \"ebb960b7-8b92-4538-a05f-d0a4460c0fc7\") " pod="openshift-authentication/oauth-openshift-76766fc778-hl47l" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.839084 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/ebb960b7-8b92-4538-a05f-d0a4460c0fc7-v4-0-config-system-router-certs\") pod \"oauth-openshift-76766fc778-hl47l\" (UID: \"ebb960b7-8b92-4538-a05f-d0a4460c0fc7\") " pod="openshift-authentication/oauth-openshift-76766fc778-hl47l" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.839579 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/ebb960b7-8b92-4538-a05f-d0a4460c0fc7-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-76766fc778-hl47l\" (UID: \"ebb960b7-8b92-4538-a05f-d0a4460c0fc7\") " pod="openshift-authentication/oauth-openshift-76766fc778-hl47l" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.839925 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/ebb960b7-8b92-4538-a05f-d0a4460c0fc7-v4-0-config-system-serving-cert\") pod \"oauth-openshift-76766fc778-hl47l\" (UID: \"ebb960b7-8b92-4538-a05f-d0a4460c0fc7\") " pod="openshift-authentication/oauth-openshift-76766fc778-hl47l" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.840413 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/ebb960b7-8b92-4538-a05f-d0a4460c0fc7-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-76766fc778-hl47l\" (UID: \"ebb960b7-8b92-4538-a05f-d0a4460c0fc7\") " pod="openshift-authentication/oauth-openshift-76766fc778-hl47l" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.840445 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/ebb960b7-8b92-4538-a05f-d0a4460c0fc7-v4-0-config-user-template-error\") pod \"oauth-openshift-76766fc778-hl47l\" (UID: \"ebb960b7-8b92-4538-a05f-d0a4460c0fc7\") " pod="openshift-authentication/oauth-openshift-76766fc778-hl47l" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.860217 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6d2w6\" (UniqueName: \"kubernetes.io/projected/ebb960b7-8b92-4538-a05f-d0a4460c0fc7-kube-api-access-6d2w6\") pod \"oauth-openshift-76766fc778-hl47l\" (UID: \"ebb960b7-8b92-4538-a05f-d0a4460c0fc7\") " pod="openshift-authentication/oauth-openshift-76766fc778-hl47l" Oct 08 00:10:27 crc kubenswrapper[4967]: I1008 00:10:27.900949 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-76766fc778-hl47l" Oct 08 00:10:28 crc kubenswrapper[4967]: I1008 00:10:28.130926 4967 generic.go:334] "Generic (PLEG): container finished" podID="d7ec4b5e-d6b4-4681-a231-c7811661c6cb" containerID="30495094bf636d17573cf31907c204329ec27e29d279878d6eeecc6249b8a486" exitCode=0 Oct 08 00:10:28 crc kubenswrapper[4967]: I1008 00:10:28.131065 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-tmj7c" Oct 08 00:10:28 crc kubenswrapper[4967]: I1008 00:10:28.139424 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-tmj7c" event={"ID":"d7ec4b5e-d6b4-4681-a231-c7811661c6cb","Type":"ContainerDied","Data":"30495094bf636d17573cf31907c204329ec27e29d279878d6eeecc6249b8a486"} Oct 08 00:10:28 crc kubenswrapper[4967]: I1008 00:10:28.139541 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-tmj7c" event={"ID":"d7ec4b5e-d6b4-4681-a231-c7811661c6cb","Type":"ContainerDied","Data":"5f315a5ae87bfac31e4598e500fd632a163a51d5846d1541929ce7ca88ed7a1b"} Oct 08 00:10:28 crc kubenswrapper[4967]: I1008 00:10:28.139592 4967 scope.go:117] "RemoveContainer" containerID="30495094bf636d17573cf31907c204329ec27e29d279878d6eeecc6249b8a486" Oct 08 00:10:28 crc kubenswrapper[4967]: I1008 00:10:28.156961 4967 scope.go:117] "RemoveContainer" containerID="30495094bf636d17573cf31907c204329ec27e29d279878d6eeecc6249b8a486" Oct 08 00:10:28 crc kubenswrapper[4967]: E1008 00:10:28.157290 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"30495094bf636d17573cf31907c204329ec27e29d279878d6eeecc6249b8a486\": container with ID starting with 30495094bf636d17573cf31907c204329ec27e29d279878d6eeecc6249b8a486 not found: ID does not exist" containerID="30495094bf636d17573cf31907c204329ec27e29d279878d6eeecc6249b8a486" Oct 08 00:10:28 crc kubenswrapper[4967]: I1008 00:10:28.157325 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"30495094bf636d17573cf31907c204329ec27e29d279878d6eeecc6249b8a486"} err="failed to get container status \"30495094bf636d17573cf31907c204329ec27e29d279878d6eeecc6249b8a486\": rpc error: code = NotFound desc = could not find container \"30495094bf636d17573cf31907c204329ec27e29d279878d6eeecc6249b8a486\": container with ID starting with 30495094bf636d17573cf31907c204329ec27e29d279878d6eeecc6249b8a486 not found: ID does not exist" Oct 08 00:10:28 crc kubenswrapper[4967]: I1008 00:10:28.174115 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-tmj7c"] Oct 08 00:10:28 crc kubenswrapper[4967]: I1008 00:10:28.177255 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-tmj7c"] Oct 08 00:10:28 crc kubenswrapper[4967]: I1008 00:10:28.400705 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-76766fc778-hl47l"] Oct 08 00:10:28 crc kubenswrapper[4967]: W1008 00:10:28.404633 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podebb960b7_8b92_4538_a05f_d0a4460c0fc7.slice/crio-24a933f2690a29bc0a47a6dc6da686634f05f3d17dc97e377da00d2a2f3e049d WatchSource:0}: Error finding container 24a933f2690a29bc0a47a6dc6da686634f05f3d17dc97e377da00d2a2f3e049d: Status 404 returned error can't find the container with id 24a933f2690a29bc0a47a6dc6da686634f05f3d17dc97e377da00d2a2f3e049d Oct 08 00:10:29 crc kubenswrapper[4967]: I1008 00:10:29.142317 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-76766fc778-hl47l" event={"ID":"ebb960b7-8b92-4538-a05f-d0a4460c0fc7","Type":"ContainerStarted","Data":"222d84e6f6978ab1d6ab6ffe51779f9dbf1e9fb81041c2c22eeccb82c795f67e"} Oct 08 00:10:29 crc kubenswrapper[4967]: I1008 00:10:29.144020 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-76766fc778-hl47l" event={"ID":"ebb960b7-8b92-4538-a05f-d0a4460c0fc7","Type":"ContainerStarted","Data":"24a933f2690a29bc0a47a6dc6da686634f05f3d17dc97e377da00d2a2f3e049d"} Oct 08 00:10:29 crc kubenswrapper[4967]: I1008 00:10:29.144206 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-76766fc778-hl47l" Oct 08 00:10:29 crc kubenswrapper[4967]: I1008 00:10:29.186133 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-76766fc778-hl47l" podStartSLOduration=27.186098877 podStartE2EDuration="27.186098877s" podCreationTimestamp="2025-10-08 00:10:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:10:29.175598139 +0000 UTC m=+227.826132818" watchObservedRunningTime="2025-10-08 00:10:29.186098877 +0000 UTC m=+227.836633566" Oct 08 00:10:29 crc kubenswrapper[4967]: I1008 00:10:29.319469 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-76766fc778-hl47l" Oct 08 00:10:30 crc kubenswrapper[4967]: I1008 00:10:30.130455 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7ec4b5e-d6b4-4681-a231-c7811661c6cb" path="/var/lib/kubelet/pods/d7ec4b5e-d6b4-4681-a231-c7811661c6cb/volumes" Oct 08 00:10:44 crc kubenswrapper[4967]: I1008 00:10:44.680897 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-x8d4p"] Oct 08 00:10:44 crc kubenswrapper[4967]: I1008 00:10:44.681863 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-x8d4p" podUID="32d2d299-2970-40e9-9116-74a074c80eb1" containerName="registry-server" containerID="cri-o://1a3ee19601a511037afb82db3574b9dbfd14a0521374a3a7993a076bbfd08ea7" gracePeriod=30 Oct 08 00:10:44 crc kubenswrapper[4967]: I1008 00:10:44.692780 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9glfc"] Oct 08 00:10:44 crc kubenswrapper[4967]: I1008 00:10:44.693063 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-9glfc" podUID="f174780f-4c64-427f-b527-91a3871c271d" containerName="registry-server" containerID="cri-o://015cf08ab330b3b2ce10e6a71da3dcae87b198a098e916db2599b8b24d67437c" gracePeriod=30 Oct 08 00:10:44 crc kubenswrapper[4967]: I1008 00:10:44.700133 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-npdfq"] Oct 08 00:10:44 crc kubenswrapper[4967]: I1008 00:10:44.700516 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-npdfq" podUID="f79e3965-8fdc-48ba-be6c-2866348f2458" containerName="marketplace-operator" containerID="cri-o://34577d89f30ca78a8888d3edc5d18f6559afb5a278eb11dd7da7f5d531cb81d2" gracePeriod=30 Oct 08 00:10:44 crc kubenswrapper[4967]: I1008 00:10:44.710423 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rzc76"] Oct 08 00:10:44 crc kubenswrapper[4967]: I1008 00:10:44.710689 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-rzc76" podUID="f0340c31-098f-4936-bd5e-4ef17c48a9f1" containerName="registry-server" containerID="cri-o://096b92263447d7226b4eaf48f8590d960262063ffc31549f9efc25a3eacd8a64" gracePeriod=30 Oct 08 00:10:44 crc kubenswrapper[4967]: I1008 00:10:44.715590 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-v7scw"] Oct 08 00:10:44 crc kubenswrapper[4967]: I1008 00:10:44.715874 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-v7scw" podUID="f3001d43-c530-4c5e-a6e6-27346682ba5c" containerName="registry-server" containerID="cri-o://4ec9fbb13bbd2a22a32c0a838bf811148fdc82af7c822a8b9a9073d9c1baf64e" gracePeriod=30 Oct 08 00:10:44 crc kubenswrapper[4967]: I1008 00:10:44.719094 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-szxjs"] Oct 08 00:10:44 crc kubenswrapper[4967]: I1008 00:10:44.719754 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-szxjs" Oct 08 00:10:44 crc kubenswrapper[4967]: I1008 00:10:44.726650 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-szxjs"] Oct 08 00:10:44 crc kubenswrapper[4967]: I1008 00:10:44.784425 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6acb8d28-53a4-46b1-93ef-5c0666234019-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-szxjs\" (UID: \"6acb8d28-53a4-46b1-93ef-5c0666234019\") " pod="openshift-marketplace/marketplace-operator-79b997595-szxjs" Oct 08 00:10:44 crc kubenswrapper[4967]: I1008 00:10:44.784505 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mkhxf\" (UniqueName: \"kubernetes.io/projected/6acb8d28-53a4-46b1-93ef-5c0666234019-kube-api-access-mkhxf\") pod \"marketplace-operator-79b997595-szxjs\" (UID: \"6acb8d28-53a4-46b1-93ef-5c0666234019\") " pod="openshift-marketplace/marketplace-operator-79b997595-szxjs" Oct 08 00:10:44 crc kubenswrapper[4967]: I1008 00:10:44.784644 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/6acb8d28-53a4-46b1-93ef-5c0666234019-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-szxjs\" (UID: \"6acb8d28-53a4-46b1-93ef-5c0666234019\") " pod="openshift-marketplace/marketplace-operator-79b997595-szxjs" Oct 08 00:10:44 crc kubenswrapper[4967]: I1008 00:10:44.845905 4967 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-npdfq container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.29:8080/healthz\": dial tcp 10.217.0.29:8080: connect: connection refused" start-of-body= Oct 08 00:10:44 crc kubenswrapper[4967]: I1008 00:10:44.845950 4967 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-npdfq" podUID="f79e3965-8fdc-48ba-be6c-2866348f2458" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.29:8080/healthz\": dial tcp 10.217.0.29:8080: connect: connection refused" Oct 08 00:10:44 crc kubenswrapper[4967]: I1008 00:10:44.886035 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mkhxf\" (UniqueName: \"kubernetes.io/projected/6acb8d28-53a4-46b1-93ef-5c0666234019-kube-api-access-mkhxf\") pod \"marketplace-operator-79b997595-szxjs\" (UID: \"6acb8d28-53a4-46b1-93ef-5c0666234019\") " pod="openshift-marketplace/marketplace-operator-79b997595-szxjs" Oct 08 00:10:44 crc kubenswrapper[4967]: I1008 00:10:44.886073 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/6acb8d28-53a4-46b1-93ef-5c0666234019-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-szxjs\" (UID: \"6acb8d28-53a4-46b1-93ef-5c0666234019\") " pod="openshift-marketplace/marketplace-operator-79b997595-szxjs" Oct 08 00:10:44 crc kubenswrapper[4967]: I1008 00:10:44.886100 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6acb8d28-53a4-46b1-93ef-5c0666234019-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-szxjs\" (UID: \"6acb8d28-53a4-46b1-93ef-5c0666234019\") " pod="openshift-marketplace/marketplace-operator-79b997595-szxjs" Oct 08 00:10:44 crc kubenswrapper[4967]: I1008 00:10:44.887058 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6acb8d28-53a4-46b1-93ef-5c0666234019-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-szxjs\" (UID: \"6acb8d28-53a4-46b1-93ef-5c0666234019\") " pod="openshift-marketplace/marketplace-operator-79b997595-szxjs" Oct 08 00:10:44 crc kubenswrapper[4967]: I1008 00:10:44.894002 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/6acb8d28-53a4-46b1-93ef-5c0666234019-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-szxjs\" (UID: \"6acb8d28-53a4-46b1-93ef-5c0666234019\") " pod="openshift-marketplace/marketplace-operator-79b997595-szxjs" Oct 08 00:10:44 crc kubenswrapper[4967]: I1008 00:10:44.902320 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mkhxf\" (UniqueName: \"kubernetes.io/projected/6acb8d28-53a4-46b1-93ef-5c0666234019-kube-api-access-mkhxf\") pod \"marketplace-operator-79b997595-szxjs\" (UID: \"6acb8d28-53a4-46b1-93ef-5c0666234019\") " pod="openshift-marketplace/marketplace-operator-79b997595-szxjs" Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.060416 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-szxjs" Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.180532 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9glfc" Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.187004 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-x8d4p" Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.191956 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-npdfq" Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.200014 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-v7scw" Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.207836 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rzc76" Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.248055 4967 generic.go:334] "Generic (PLEG): container finished" podID="f79e3965-8fdc-48ba-be6c-2866348f2458" containerID="34577d89f30ca78a8888d3edc5d18f6559afb5a278eb11dd7da7f5d531cb81d2" exitCode=0 Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.248121 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-npdfq" Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.248268 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-npdfq" event={"ID":"f79e3965-8fdc-48ba-be6c-2866348f2458","Type":"ContainerDied","Data":"34577d89f30ca78a8888d3edc5d18f6559afb5a278eb11dd7da7f5d531cb81d2"} Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.248406 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-npdfq" event={"ID":"f79e3965-8fdc-48ba-be6c-2866348f2458","Type":"ContainerDied","Data":"5606f3698d1288dfa61fe48f4444791a2fcc5155f6c140d0adb486c975c97fa1"} Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.248443 4967 scope.go:117] "RemoveContainer" containerID="34577d89f30ca78a8888d3edc5d18f6559afb5a278eb11dd7da7f5d531cb81d2" Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.251202 4967 generic.go:334] "Generic (PLEG): container finished" podID="32d2d299-2970-40e9-9116-74a074c80eb1" containerID="1a3ee19601a511037afb82db3574b9dbfd14a0521374a3a7993a076bbfd08ea7" exitCode=0 Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.251326 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x8d4p" event={"ID":"32d2d299-2970-40e9-9116-74a074c80eb1","Type":"ContainerDied","Data":"1a3ee19601a511037afb82db3574b9dbfd14a0521374a3a7993a076bbfd08ea7"} Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.251355 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x8d4p" event={"ID":"32d2d299-2970-40e9-9116-74a074c80eb1","Type":"ContainerDied","Data":"4d7207b9b44c47786d92cec91a19c4ddf87acaa341e7d59d18393732ed3ef1b0"} Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.251467 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-x8d4p" Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.266413 4967 generic.go:334] "Generic (PLEG): container finished" podID="f0340c31-098f-4936-bd5e-4ef17c48a9f1" containerID="096b92263447d7226b4eaf48f8590d960262063ffc31549f9efc25a3eacd8a64" exitCode=0 Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.266499 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rzc76" event={"ID":"f0340c31-098f-4936-bd5e-4ef17c48a9f1","Type":"ContainerDied","Data":"096b92263447d7226b4eaf48f8590d960262063ffc31549f9efc25a3eacd8a64"} Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.266505 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rzc76" Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.266530 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rzc76" event={"ID":"f0340c31-098f-4936-bd5e-4ef17c48a9f1","Type":"ContainerDied","Data":"cf7ef634599a2521aeb21dcf825d637113880f802e35fa8ebeeed633ffb7d906"} Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.278944 4967 generic.go:334] "Generic (PLEG): container finished" podID="f174780f-4c64-427f-b527-91a3871c271d" containerID="015cf08ab330b3b2ce10e6a71da3dcae87b198a098e916db2599b8b24d67437c" exitCode=0 Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.278986 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9glfc" event={"ID":"f174780f-4c64-427f-b527-91a3871c271d","Type":"ContainerDied","Data":"015cf08ab330b3b2ce10e6a71da3dcae87b198a098e916db2599b8b24d67437c"} Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.279026 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9glfc" event={"ID":"f174780f-4c64-427f-b527-91a3871c271d","Type":"ContainerDied","Data":"2a89b55c2d7c99f1995c42994d2f74ab6d2ad5cbb2a4bda1d5c74f9636741158"} Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.279051 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9glfc" Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.281531 4967 generic.go:334] "Generic (PLEG): container finished" podID="f3001d43-c530-4c5e-a6e6-27346682ba5c" containerID="4ec9fbb13bbd2a22a32c0a838bf811148fdc82af7c822a8b9a9073d9c1baf64e" exitCode=0 Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.281557 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v7scw" event={"ID":"f3001d43-c530-4c5e-a6e6-27346682ba5c","Type":"ContainerDied","Data":"4ec9fbb13bbd2a22a32c0a838bf811148fdc82af7c822a8b9a9073d9c1baf64e"} Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.281574 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v7scw" event={"ID":"f3001d43-c530-4c5e-a6e6-27346682ba5c","Type":"ContainerDied","Data":"79f4a8d4284a11c7df546ef08ef87f4a94b04fe0b759ceb5097711008c24bcfa"} Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.281624 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-v7scw" Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.284820 4967 scope.go:117] "RemoveContainer" containerID="34577d89f30ca78a8888d3edc5d18f6559afb5a278eb11dd7da7f5d531cb81d2" Oct 08 00:10:45 crc kubenswrapper[4967]: E1008 00:10:45.285435 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"34577d89f30ca78a8888d3edc5d18f6559afb5a278eb11dd7da7f5d531cb81d2\": container with ID starting with 34577d89f30ca78a8888d3edc5d18f6559afb5a278eb11dd7da7f5d531cb81d2 not found: ID does not exist" containerID="34577d89f30ca78a8888d3edc5d18f6559afb5a278eb11dd7da7f5d531cb81d2" Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.285464 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"34577d89f30ca78a8888d3edc5d18f6559afb5a278eb11dd7da7f5d531cb81d2"} err="failed to get container status \"34577d89f30ca78a8888d3edc5d18f6559afb5a278eb11dd7da7f5d531cb81d2\": rpc error: code = NotFound desc = could not find container \"34577d89f30ca78a8888d3edc5d18f6559afb5a278eb11dd7da7f5d531cb81d2\": container with ID starting with 34577d89f30ca78a8888d3edc5d18f6559afb5a278eb11dd7da7f5d531cb81d2 not found: ID does not exist" Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.285491 4967 scope.go:117] "RemoveContainer" containerID="1a3ee19601a511037afb82db3574b9dbfd14a0521374a3a7993a076bbfd08ea7" Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.293515 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f174780f-4c64-427f-b527-91a3871c271d-utilities\") pod \"f174780f-4c64-427f-b527-91a3871c271d\" (UID: \"f174780f-4c64-427f-b527-91a3871c271d\") " Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.293550 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f174780f-4c64-427f-b527-91a3871c271d-catalog-content\") pod \"f174780f-4c64-427f-b527-91a3871c271d\" (UID: \"f174780f-4c64-427f-b527-91a3871c271d\") " Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.293579 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/32d2d299-2970-40e9-9116-74a074c80eb1-catalog-content\") pod \"32d2d299-2970-40e9-9116-74a074c80eb1\" (UID: \"32d2d299-2970-40e9-9116-74a074c80eb1\") " Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.293625 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w22ps\" (UniqueName: \"kubernetes.io/projected/f79e3965-8fdc-48ba-be6c-2866348f2458-kube-api-access-w22ps\") pod \"f79e3965-8fdc-48ba-be6c-2866348f2458\" (UID: \"f79e3965-8fdc-48ba-be6c-2866348f2458\") " Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.293661 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-775qd\" (UniqueName: \"kubernetes.io/projected/32d2d299-2970-40e9-9116-74a074c80eb1-kube-api-access-775qd\") pod \"32d2d299-2970-40e9-9116-74a074c80eb1\" (UID: \"32d2d299-2970-40e9-9116-74a074c80eb1\") " Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.293908 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f79e3965-8fdc-48ba-be6c-2866348f2458-marketplace-operator-metrics\") pod \"f79e3965-8fdc-48ba-be6c-2866348f2458\" (UID: \"f79e3965-8fdc-48ba-be6c-2866348f2458\") " Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.293961 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/32d2d299-2970-40e9-9116-74a074c80eb1-utilities\") pod \"32d2d299-2970-40e9-9116-74a074c80eb1\" (UID: \"32d2d299-2970-40e9-9116-74a074c80eb1\") " Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.293995 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-blhnw\" (UniqueName: \"kubernetes.io/projected/f174780f-4c64-427f-b527-91a3871c271d-kube-api-access-blhnw\") pod \"f174780f-4c64-427f-b527-91a3871c271d\" (UID: \"f174780f-4c64-427f-b527-91a3871c271d\") " Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.294034 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3001d43-c530-4c5e-a6e6-27346682ba5c-catalog-content\") pod \"f3001d43-c530-4c5e-a6e6-27346682ba5c\" (UID: \"f3001d43-c530-4c5e-a6e6-27346682ba5c\") " Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.294058 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f79e3965-8fdc-48ba-be6c-2866348f2458-marketplace-trusted-ca\") pod \"f79e3965-8fdc-48ba-be6c-2866348f2458\" (UID: \"f79e3965-8fdc-48ba-be6c-2866348f2458\") " Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.295073 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f79e3965-8fdc-48ba-be6c-2866348f2458-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "f79e3965-8fdc-48ba-be6c-2866348f2458" (UID: "f79e3965-8fdc-48ba-be6c-2866348f2458"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.295172 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f174780f-4c64-427f-b527-91a3871c271d-utilities" (OuterVolumeSpecName: "utilities") pod "f174780f-4c64-427f-b527-91a3871c271d" (UID: "f174780f-4c64-427f-b527-91a3871c271d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.295552 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/32d2d299-2970-40e9-9116-74a074c80eb1-utilities" (OuterVolumeSpecName: "utilities") pod "32d2d299-2970-40e9-9116-74a074c80eb1" (UID: "32d2d299-2970-40e9-9116-74a074c80eb1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.297869 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32d2d299-2970-40e9-9116-74a074c80eb1-kube-api-access-775qd" (OuterVolumeSpecName: "kube-api-access-775qd") pod "32d2d299-2970-40e9-9116-74a074c80eb1" (UID: "32d2d299-2970-40e9-9116-74a074c80eb1"). InnerVolumeSpecName "kube-api-access-775qd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.298239 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f79e3965-8fdc-48ba-be6c-2866348f2458-kube-api-access-w22ps" (OuterVolumeSpecName: "kube-api-access-w22ps") pod "f79e3965-8fdc-48ba-be6c-2866348f2458" (UID: "f79e3965-8fdc-48ba-be6c-2866348f2458"). InnerVolumeSpecName "kube-api-access-w22ps". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.298574 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f79e3965-8fdc-48ba-be6c-2866348f2458-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "f79e3965-8fdc-48ba-be6c-2866348f2458" (UID: "f79e3965-8fdc-48ba-be6c-2866348f2458"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.298612 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f174780f-4c64-427f-b527-91a3871c271d-kube-api-access-blhnw" (OuterVolumeSpecName: "kube-api-access-blhnw") pod "f174780f-4c64-427f-b527-91a3871c271d" (UID: "f174780f-4c64-427f-b527-91a3871c271d"). InnerVolumeSpecName "kube-api-access-blhnw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.302176 4967 scope.go:117] "RemoveContainer" containerID="1430349593ed2e61050f2eb745cfad0b82ef9f6e459baee55b2d54a8c2c11593" Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.321719 4967 scope.go:117] "RemoveContainer" containerID="0425a89b764948462abab2813a5e0a67578e7f549efe81290c8da7aaa8337561" Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.338681 4967 scope.go:117] "RemoveContainer" containerID="1a3ee19601a511037afb82db3574b9dbfd14a0521374a3a7993a076bbfd08ea7" Oct 08 00:10:45 crc kubenswrapper[4967]: E1008 00:10:45.339530 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a3ee19601a511037afb82db3574b9dbfd14a0521374a3a7993a076bbfd08ea7\": container with ID starting with 1a3ee19601a511037afb82db3574b9dbfd14a0521374a3a7993a076bbfd08ea7 not found: ID does not exist" containerID="1a3ee19601a511037afb82db3574b9dbfd14a0521374a3a7993a076bbfd08ea7" Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.339578 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a3ee19601a511037afb82db3574b9dbfd14a0521374a3a7993a076bbfd08ea7"} err="failed to get container status \"1a3ee19601a511037afb82db3574b9dbfd14a0521374a3a7993a076bbfd08ea7\": rpc error: code = NotFound desc = could not find container \"1a3ee19601a511037afb82db3574b9dbfd14a0521374a3a7993a076bbfd08ea7\": container with ID starting with 1a3ee19601a511037afb82db3574b9dbfd14a0521374a3a7993a076bbfd08ea7 not found: ID does not exist" Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.339606 4967 scope.go:117] "RemoveContainer" containerID="1430349593ed2e61050f2eb745cfad0b82ef9f6e459baee55b2d54a8c2c11593" Oct 08 00:10:45 crc kubenswrapper[4967]: E1008 00:10:45.340040 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1430349593ed2e61050f2eb745cfad0b82ef9f6e459baee55b2d54a8c2c11593\": container with ID starting with 1430349593ed2e61050f2eb745cfad0b82ef9f6e459baee55b2d54a8c2c11593 not found: ID does not exist" containerID="1430349593ed2e61050f2eb745cfad0b82ef9f6e459baee55b2d54a8c2c11593" Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.340103 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1430349593ed2e61050f2eb745cfad0b82ef9f6e459baee55b2d54a8c2c11593"} err="failed to get container status \"1430349593ed2e61050f2eb745cfad0b82ef9f6e459baee55b2d54a8c2c11593\": rpc error: code = NotFound desc = could not find container \"1430349593ed2e61050f2eb745cfad0b82ef9f6e459baee55b2d54a8c2c11593\": container with ID starting with 1430349593ed2e61050f2eb745cfad0b82ef9f6e459baee55b2d54a8c2c11593 not found: ID does not exist" Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.340138 4967 scope.go:117] "RemoveContainer" containerID="0425a89b764948462abab2813a5e0a67578e7f549efe81290c8da7aaa8337561" Oct 08 00:10:45 crc kubenswrapper[4967]: E1008 00:10:45.340405 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0425a89b764948462abab2813a5e0a67578e7f549efe81290c8da7aaa8337561\": container with ID starting with 0425a89b764948462abab2813a5e0a67578e7f549efe81290c8da7aaa8337561 not found: ID does not exist" containerID="0425a89b764948462abab2813a5e0a67578e7f549efe81290c8da7aaa8337561" Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.340444 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0425a89b764948462abab2813a5e0a67578e7f549efe81290c8da7aaa8337561"} err="failed to get container status \"0425a89b764948462abab2813a5e0a67578e7f549efe81290c8da7aaa8337561\": rpc error: code = NotFound desc = could not find container \"0425a89b764948462abab2813a5e0a67578e7f549efe81290c8da7aaa8337561\": container with ID starting with 0425a89b764948462abab2813a5e0a67578e7f549efe81290c8da7aaa8337561 not found: ID does not exist" Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.340468 4967 scope.go:117] "RemoveContainer" containerID="096b92263447d7226b4eaf48f8590d960262063ffc31549f9efc25a3eacd8a64" Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.357004 4967 scope.go:117] "RemoveContainer" containerID="22507f8b5ef6f57835f248285e841020cdaf6f9a17f62e5ec6a75d3bdeaaa127" Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.364849 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f174780f-4c64-427f-b527-91a3871c271d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f174780f-4c64-427f-b527-91a3871c271d" (UID: "f174780f-4c64-427f-b527-91a3871c271d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.370721 4967 scope.go:117] "RemoveContainer" containerID="6e8c3a1ca8260e2aff066ffa90cac9d502bed952d2d7901effc492dabd1c3702" Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.371990 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/32d2d299-2970-40e9-9116-74a074c80eb1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "32d2d299-2970-40e9-9116-74a074c80eb1" (UID: "32d2d299-2970-40e9-9116-74a074c80eb1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.395035 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3001d43-c530-4c5e-a6e6-27346682ba5c-utilities\") pod \"f3001d43-c530-4c5e-a6e6-27346682ba5c\" (UID: \"f3001d43-c530-4c5e-a6e6-27346682ba5c\") " Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.395094 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f0340c31-098f-4936-bd5e-4ef17c48a9f1-catalog-content\") pod \"f0340c31-098f-4936-bd5e-4ef17c48a9f1\" (UID: \"f0340c31-098f-4936-bd5e-4ef17c48a9f1\") " Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.395142 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f0340c31-098f-4936-bd5e-4ef17c48a9f1-utilities\") pod \"f0340c31-098f-4936-bd5e-4ef17c48a9f1\" (UID: \"f0340c31-098f-4936-bd5e-4ef17c48a9f1\") " Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.395244 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5qq8z\" (UniqueName: \"kubernetes.io/projected/f3001d43-c530-4c5e-a6e6-27346682ba5c-kube-api-access-5qq8z\") pod \"f3001d43-c530-4c5e-a6e6-27346682ba5c\" (UID: \"f3001d43-c530-4c5e-a6e6-27346682ba5c\") " Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.395279 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fft4x\" (UniqueName: \"kubernetes.io/projected/f0340c31-098f-4936-bd5e-4ef17c48a9f1-kube-api-access-fft4x\") pod \"f0340c31-098f-4936-bd5e-4ef17c48a9f1\" (UID: \"f0340c31-098f-4936-bd5e-4ef17c48a9f1\") " Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.395523 4967 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f174780f-4c64-427f-b527-91a3871c271d-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.395546 4967 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f174780f-4c64-427f-b527-91a3871c271d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.395558 4967 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/32d2d299-2970-40e9-9116-74a074c80eb1-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.395573 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w22ps\" (UniqueName: \"kubernetes.io/projected/f79e3965-8fdc-48ba-be6c-2866348f2458-kube-api-access-w22ps\") on node \"crc\" DevicePath \"\"" Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.395585 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-775qd\" (UniqueName: \"kubernetes.io/projected/32d2d299-2970-40e9-9116-74a074c80eb1-kube-api-access-775qd\") on node \"crc\" DevicePath \"\"" Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.395598 4967 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/f79e3965-8fdc-48ba-be6c-2866348f2458-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.395610 4967 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/32d2d299-2970-40e9-9116-74a074c80eb1-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.395622 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-blhnw\" (UniqueName: \"kubernetes.io/projected/f174780f-4c64-427f-b527-91a3871c271d-kube-api-access-blhnw\") on node \"crc\" DevicePath \"\"" Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.395634 4967 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f79e3965-8fdc-48ba-be6c-2866348f2458-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.395845 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f3001d43-c530-4c5e-a6e6-27346682ba5c-utilities" (OuterVolumeSpecName: "utilities") pod "f3001d43-c530-4c5e-a6e6-27346682ba5c" (UID: "f3001d43-c530-4c5e-a6e6-27346682ba5c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.396122 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f0340c31-098f-4936-bd5e-4ef17c48a9f1-utilities" (OuterVolumeSpecName: "utilities") pod "f0340c31-098f-4936-bd5e-4ef17c48a9f1" (UID: "f0340c31-098f-4936-bd5e-4ef17c48a9f1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.396407 4967 scope.go:117] "RemoveContainer" containerID="096b92263447d7226b4eaf48f8590d960262063ffc31549f9efc25a3eacd8a64" Oct 08 00:10:45 crc kubenswrapper[4967]: E1008 00:10:45.396856 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"096b92263447d7226b4eaf48f8590d960262063ffc31549f9efc25a3eacd8a64\": container with ID starting with 096b92263447d7226b4eaf48f8590d960262063ffc31549f9efc25a3eacd8a64 not found: ID does not exist" containerID="096b92263447d7226b4eaf48f8590d960262063ffc31549f9efc25a3eacd8a64" Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.396886 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"096b92263447d7226b4eaf48f8590d960262063ffc31549f9efc25a3eacd8a64"} err="failed to get container status \"096b92263447d7226b4eaf48f8590d960262063ffc31549f9efc25a3eacd8a64\": rpc error: code = NotFound desc = could not find container \"096b92263447d7226b4eaf48f8590d960262063ffc31549f9efc25a3eacd8a64\": container with ID starting with 096b92263447d7226b4eaf48f8590d960262063ffc31549f9efc25a3eacd8a64 not found: ID does not exist" Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.396914 4967 scope.go:117] "RemoveContainer" containerID="22507f8b5ef6f57835f248285e841020cdaf6f9a17f62e5ec6a75d3bdeaaa127" Oct 08 00:10:45 crc kubenswrapper[4967]: E1008 00:10:45.397136 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"22507f8b5ef6f57835f248285e841020cdaf6f9a17f62e5ec6a75d3bdeaaa127\": container with ID starting with 22507f8b5ef6f57835f248285e841020cdaf6f9a17f62e5ec6a75d3bdeaaa127 not found: ID does not exist" containerID="22507f8b5ef6f57835f248285e841020cdaf6f9a17f62e5ec6a75d3bdeaaa127" Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.397160 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"22507f8b5ef6f57835f248285e841020cdaf6f9a17f62e5ec6a75d3bdeaaa127"} err="failed to get container status \"22507f8b5ef6f57835f248285e841020cdaf6f9a17f62e5ec6a75d3bdeaaa127\": rpc error: code = NotFound desc = could not find container \"22507f8b5ef6f57835f248285e841020cdaf6f9a17f62e5ec6a75d3bdeaaa127\": container with ID starting with 22507f8b5ef6f57835f248285e841020cdaf6f9a17f62e5ec6a75d3bdeaaa127 not found: ID does not exist" Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.397177 4967 scope.go:117] "RemoveContainer" containerID="6e8c3a1ca8260e2aff066ffa90cac9d502bed952d2d7901effc492dabd1c3702" Oct 08 00:10:45 crc kubenswrapper[4967]: E1008 00:10:45.397406 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6e8c3a1ca8260e2aff066ffa90cac9d502bed952d2d7901effc492dabd1c3702\": container with ID starting with 6e8c3a1ca8260e2aff066ffa90cac9d502bed952d2d7901effc492dabd1c3702 not found: ID does not exist" containerID="6e8c3a1ca8260e2aff066ffa90cac9d502bed952d2d7901effc492dabd1c3702" Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.397428 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e8c3a1ca8260e2aff066ffa90cac9d502bed952d2d7901effc492dabd1c3702"} err="failed to get container status \"6e8c3a1ca8260e2aff066ffa90cac9d502bed952d2d7901effc492dabd1c3702\": rpc error: code = NotFound desc = could not find container \"6e8c3a1ca8260e2aff066ffa90cac9d502bed952d2d7901effc492dabd1c3702\": container with ID starting with 6e8c3a1ca8260e2aff066ffa90cac9d502bed952d2d7901effc492dabd1c3702 not found: ID does not exist" Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.397444 4967 scope.go:117] "RemoveContainer" containerID="015cf08ab330b3b2ce10e6a71da3dcae87b198a098e916db2599b8b24d67437c" Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.399380 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0340c31-098f-4936-bd5e-4ef17c48a9f1-kube-api-access-fft4x" (OuterVolumeSpecName: "kube-api-access-fft4x") pod "f0340c31-098f-4936-bd5e-4ef17c48a9f1" (UID: "f0340c31-098f-4936-bd5e-4ef17c48a9f1"). InnerVolumeSpecName "kube-api-access-fft4x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.399457 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3001d43-c530-4c5e-a6e6-27346682ba5c-kube-api-access-5qq8z" (OuterVolumeSpecName: "kube-api-access-5qq8z") pod "f3001d43-c530-4c5e-a6e6-27346682ba5c" (UID: "f3001d43-c530-4c5e-a6e6-27346682ba5c"). InnerVolumeSpecName "kube-api-access-5qq8z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.409772 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f0340c31-098f-4936-bd5e-4ef17c48a9f1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f0340c31-098f-4936-bd5e-4ef17c48a9f1" (UID: "f0340c31-098f-4936-bd5e-4ef17c48a9f1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.410407 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f3001d43-c530-4c5e-a6e6-27346682ba5c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f3001d43-c530-4c5e-a6e6-27346682ba5c" (UID: "f3001d43-c530-4c5e-a6e6-27346682ba5c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.414987 4967 scope.go:117] "RemoveContainer" containerID="f016083cd27bea7cf6bf1f296c6f9468ea1645fffa309886dcf8a3e9bfd1e5df" Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.432793 4967 scope.go:117] "RemoveContainer" containerID="e242ae473cda65bdf80d16e7b1ae127906db7b7f307e0b71b636d1caa3f3818c" Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.459536 4967 scope.go:117] "RemoveContainer" containerID="015cf08ab330b3b2ce10e6a71da3dcae87b198a098e916db2599b8b24d67437c" Oct 08 00:10:45 crc kubenswrapper[4967]: E1008 00:10:45.460103 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"015cf08ab330b3b2ce10e6a71da3dcae87b198a098e916db2599b8b24d67437c\": container with ID starting with 015cf08ab330b3b2ce10e6a71da3dcae87b198a098e916db2599b8b24d67437c not found: ID does not exist" containerID="015cf08ab330b3b2ce10e6a71da3dcae87b198a098e916db2599b8b24d67437c" Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.460147 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"015cf08ab330b3b2ce10e6a71da3dcae87b198a098e916db2599b8b24d67437c"} err="failed to get container status \"015cf08ab330b3b2ce10e6a71da3dcae87b198a098e916db2599b8b24d67437c\": rpc error: code = NotFound desc = could not find container \"015cf08ab330b3b2ce10e6a71da3dcae87b198a098e916db2599b8b24d67437c\": container with ID starting with 015cf08ab330b3b2ce10e6a71da3dcae87b198a098e916db2599b8b24d67437c not found: ID does not exist" Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.460182 4967 scope.go:117] "RemoveContainer" containerID="f016083cd27bea7cf6bf1f296c6f9468ea1645fffa309886dcf8a3e9bfd1e5df" Oct 08 00:10:45 crc kubenswrapper[4967]: E1008 00:10:45.460913 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f016083cd27bea7cf6bf1f296c6f9468ea1645fffa309886dcf8a3e9bfd1e5df\": container with ID starting with f016083cd27bea7cf6bf1f296c6f9468ea1645fffa309886dcf8a3e9bfd1e5df not found: ID does not exist" containerID="f016083cd27bea7cf6bf1f296c6f9468ea1645fffa309886dcf8a3e9bfd1e5df" Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.460979 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f016083cd27bea7cf6bf1f296c6f9468ea1645fffa309886dcf8a3e9bfd1e5df"} err="failed to get container status \"f016083cd27bea7cf6bf1f296c6f9468ea1645fffa309886dcf8a3e9bfd1e5df\": rpc error: code = NotFound desc = could not find container \"f016083cd27bea7cf6bf1f296c6f9468ea1645fffa309886dcf8a3e9bfd1e5df\": container with ID starting with f016083cd27bea7cf6bf1f296c6f9468ea1645fffa309886dcf8a3e9bfd1e5df not found: ID does not exist" Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.461042 4967 scope.go:117] "RemoveContainer" containerID="e242ae473cda65bdf80d16e7b1ae127906db7b7f307e0b71b636d1caa3f3818c" Oct 08 00:10:45 crc kubenswrapper[4967]: E1008 00:10:45.463927 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e242ae473cda65bdf80d16e7b1ae127906db7b7f307e0b71b636d1caa3f3818c\": container with ID starting with e242ae473cda65bdf80d16e7b1ae127906db7b7f307e0b71b636d1caa3f3818c not found: ID does not exist" containerID="e242ae473cda65bdf80d16e7b1ae127906db7b7f307e0b71b636d1caa3f3818c" Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.463973 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e242ae473cda65bdf80d16e7b1ae127906db7b7f307e0b71b636d1caa3f3818c"} err="failed to get container status \"e242ae473cda65bdf80d16e7b1ae127906db7b7f307e0b71b636d1caa3f3818c\": rpc error: code = NotFound desc = could not find container \"e242ae473cda65bdf80d16e7b1ae127906db7b7f307e0b71b636d1caa3f3818c\": container with ID starting with e242ae473cda65bdf80d16e7b1ae127906db7b7f307e0b71b636d1caa3f3818c not found: ID does not exist" Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.464003 4967 scope.go:117] "RemoveContainer" containerID="4ec9fbb13bbd2a22a32c0a838bf811148fdc82af7c822a8b9a9073d9c1baf64e" Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.467826 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-szxjs"] Oct 08 00:10:45 crc kubenswrapper[4967]: W1008 00:10:45.477998 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6acb8d28_53a4_46b1_93ef_5c0666234019.slice/crio-70e9f3cb830b643d2acadca7a85d709ddc663c38057f8b2c854333f819352680 WatchSource:0}: Error finding container 70e9f3cb830b643d2acadca7a85d709ddc663c38057f8b2c854333f819352680: Status 404 returned error can't find the container with id 70e9f3cb830b643d2acadca7a85d709ddc663c38057f8b2c854333f819352680 Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.478348 4967 scope.go:117] "RemoveContainer" containerID="454fcde55a49691a74ac4ad160859f26321f0cf958740208ac85065d2a653837" Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.496921 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5qq8z\" (UniqueName: \"kubernetes.io/projected/f3001d43-c530-4c5e-a6e6-27346682ba5c-kube-api-access-5qq8z\") on node \"crc\" DevicePath \"\"" Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.496952 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fft4x\" (UniqueName: \"kubernetes.io/projected/f0340c31-098f-4936-bd5e-4ef17c48a9f1-kube-api-access-fft4x\") on node \"crc\" DevicePath \"\"" Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.496964 4967 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f3001d43-c530-4c5e-a6e6-27346682ba5c-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.496976 4967 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f0340c31-098f-4936-bd5e-4ef17c48a9f1-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.496987 4967 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f3001d43-c530-4c5e-a6e6-27346682ba5c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.496997 4967 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f0340c31-098f-4936-bd5e-4ef17c48a9f1-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.500596 4967 scope.go:117] "RemoveContainer" containerID="465c13717c0d1dc7774b70e329714c01d9a30e91684ef10a3fa5ffa232423eab" Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.516023 4967 scope.go:117] "RemoveContainer" containerID="4ec9fbb13bbd2a22a32c0a838bf811148fdc82af7c822a8b9a9073d9c1baf64e" Oct 08 00:10:45 crc kubenswrapper[4967]: E1008 00:10:45.516487 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4ec9fbb13bbd2a22a32c0a838bf811148fdc82af7c822a8b9a9073d9c1baf64e\": container with ID starting with 4ec9fbb13bbd2a22a32c0a838bf811148fdc82af7c822a8b9a9073d9c1baf64e not found: ID does not exist" containerID="4ec9fbb13bbd2a22a32c0a838bf811148fdc82af7c822a8b9a9073d9c1baf64e" Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.516527 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ec9fbb13bbd2a22a32c0a838bf811148fdc82af7c822a8b9a9073d9c1baf64e"} err="failed to get container status \"4ec9fbb13bbd2a22a32c0a838bf811148fdc82af7c822a8b9a9073d9c1baf64e\": rpc error: code = NotFound desc = could not find container \"4ec9fbb13bbd2a22a32c0a838bf811148fdc82af7c822a8b9a9073d9c1baf64e\": container with ID starting with 4ec9fbb13bbd2a22a32c0a838bf811148fdc82af7c822a8b9a9073d9c1baf64e not found: ID does not exist" Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.516557 4967 scope.go:117] "RemoveContainer" containerID="454fcde55a49691a74ac4ad160859f26321f0cf958740208ac85065d2a653837" Oct 08 00:10:45 crc kubenswrapper[4967]: E1008 00:10:45.516949 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"454fcde55a49691a74ac4ad160859f26321f0cf958740208ac85065d2a653837\": container with ID starting with 454fcde55a49691a74ac4ad160859f26321f0cf958740208ac85065d2a653837 not found: ID does not exist" containerID="454fcde55a49691a74ac4ad160859f26321f0cf958740208ac85065d2a653837" Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.516977 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"454fcde55a49691a74ac4ad160859f26321f0cf958740208ac85065d2a653837"} err="failed to get container status \"454fcde55a49691a74ac4ad160859f26321f0cf958740208ac85065d2a653837\": rpc error: code = NotFound desc = could not find container \"454fcde55a49691a74ac4ad160859f26321f0cf958740208ac85065d2a653837\": container with ID starting with 454fcde55a49691a74ac4ad160859f26321f0cf958740208ac85065d2a653837 not found: ID does not exist" Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.517018 4967 scope.go:117] "RemoveContainer" containerID="465c13717c0d1dc7774b70e329714c01d9a30e91684ef10a3fa5ffa232423eab" Oct 08 00:10:45 crc kubenswrapper[4967]: E1008 00:10:45.517267 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"465c13717c0d1dc7774b70e329714c01d9a30e91684ef10a3fa5ffa232423eab\": container with ID starting with 465c13717c0d1dc7774b70e329714c01d9a30e91684ef10a3fa5ffa232423eab not found: ID does not exist" containerID="465c13717c0d1dc7774b70e329714c01d9a30e91684ef10a3fa5ffa232423eab" Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.517292 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"465c13717c0d1dc7774b70e329714c01d9a30e91684ef10a3fa5ffa232423eab"} err="failed to get container status \"465c13717c0d1dc7774b70e329714c01d9a30e91684ef10a3fa5ffa232423eab\": rpc error: code = NotFound desc = could not find container \"465c13717c0d1dc7774b70e329714c01d9a30e91684ef10a3fa5ffa232423eab\": container with ID starting with 465c13717c0d1dc7774b70e329714c01d9a30e91684ef10a3fa5ffa232423eab not found: ID does not exist" Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.579254 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-npdfq"] Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.584005 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-npdfq"] Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.599304 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-x8d4p"] Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.606003 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-x8d4p"] Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.624203 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rzc76"] Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.639846 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-rzc76"] Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.646410 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9glfc"] Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.652493 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-9glfc"] Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.658346 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-v7scw"] Oct 08 00:10:45 crc kubenswrapper[4967]: I1008 00:10:45.660851 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-v7scw"] Oct 08 00:10:46 crc kubenswrapper[4967]: I1008 00:10:46.125775 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32d2d299-2970-40e9-9116-74a074c80eb1" path="/var/lib/kubelet/pods/32d2d299-2970-40e9-9116-74a074c80eb1/volumes" Oct 08 00:10:46 crc kubenswrapper[4967]: I1008 00:10:46.126355 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f0340c31-098f-4936-bd5e-4ef17c48a9f1" path="/var/lib/kubelet/pods/f0340c31-098f-4936-bd5e-4ef17c48a9f1/volumes" Oct 08 00:10:46 crc kubenswrapper[4967]: I1008 00:10:46.126944 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f174780f-4c64-427f-b527-91a3871c271d" path="/var/lib/kubelet/pods/f174780f-4c64-427f-b527-91a3871c271d/volumes" Oct 08 00:10:46 crc kubenswrapper[4967]: I1008 00:10:46.127880 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f3001d43-c530-4c5e-a6e6-27346682ba5c" path="/var/lib/kubelet/pods/f3001d43-c530-4c5e-a6e6-27346682ba5c/volumes" Oct 08 00:10:46 crc kubenswrapper[4967]: I1008 00:10:46.128431 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f79e3965-8fdc-48ba-be6c-2866348f2458" path="/var/lib/kubelet/pods/f79e3965-8fdc-48ba-be6c-2866348f2458/volumes" Oct 08 00:10:46 crc kubenswrapper[4967]: I1008 00:10:46.290956 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-szxjs" event={"ID":"6acb8d28-53a4-46b1-93ef-5c0666234019","Type":"ContainerStarted","Data":"be2fb75b4db944fe412c4b50f7b41b9666c42134274546a90c88396841f6f0f9"} Oct 08 00:10:46 crc kubenswrapper[4967]: I1008 00:10:46.290992 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-szxjs" event={"ID":"6acb8d28-53a4-46b1-93ef-5c0666234019","Type":"ContainerStarted","Data":"70e9f3cb830b643d2acadca7a85d709ddc663c38057f8b2c854333f819352680"} Oct 08 00:10:46 crc kubenswrapper[4967]: I1008 00:10:46.291102 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-szxjs" Oct 08 00:10:46 crc kubenswrapper[4967]: I1008 00:10:46.296964 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-szxjs" Oct 08 00:10:46 crc kubenswrapper[4967]: I1008 00:10:46.311035 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-szxjs" podStartSLOduration=2.311014021 podStartE2EDuration="2.311014021s" podCreationTimestamp="2025-10-08 00:10:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:10:46.306511357 +0000 UTC m=+244.957046006" watchObservedRunningTime="2025-10-08 00:10:46.311014021 +0000 UTC m=+244.961548670" Oct 08 00:10:46 crc kubenswrapper[4967]: I1008 00:10:46.666904 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-pjrvj"] Oct 08 00:10:46 crc kubenswrapper[4967]: E1008 00:10:46.667123 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f174780f-4c64-427f-b527-91a3871c271d" containerName="extract-content" Oct 08 00:10:46 crc kubenswrapper[4967]: I1008 00:10:46.667138 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="f174780f-4c64-427f-b527-91a3871c271d" containerName="extract-content" Oct 08 00:10:46 crc kubenswrapper[4967]: E1008 00:10:46.667152 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0340c31-098f-4936-bd5e-4ef17c48a9f1" containerName="extract-content" Oct 08 00:10:46 crc kubenswrapper[4967]: I1008 00:10:46.667160 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0340c31-098f-4936-bd5e-4ef17c48a9f1" containerName="extract-content" Oct 08 00:10:46 crc kubenswrapper[4967]: E1008 00:10:46.667171 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32d2d299-2970-40e9-9116-74a074c80eb1" containerName="extract-utilities" Oct 08 00:10:46 crc kubenswrapper[4967]: I1008 00:10:46.667178 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="32d2d299-2970-40e9-9116-74a074c80eb1" containerName="extract-utilities" Oct 08 00:10:46 crc kubenswrapper[4967]: E1008 00:10:46.667191 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3001d43-c530-4c5e-a6e6-27346682ba5c" containerName="extract-utilities" Oct 08 00:10:46 crc kubenswrapper[4967]: I1008 00:10:46.667200 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3001d43-c530-4c5e-a6e6-27346682ba5c" containerName="extract-utilities" Oct 08 00:10:46 crc kubenswrapper[4967]: E1008 00:10:46.667211 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f174780f-4c64-427f-b527-91a3871c271d" containerName="registry-server" Oct 08 00:10:46 crc kubenswrapper[4967]: I1008 00:10:46.667219 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="f174780f-4c64-427f-b527-91a3871c271d" containerName="registry-server" Oct 08 00:10:46 crc kubenswrapper[4967]: E1008 00:10:46.667228 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f174780f-4c64-427f-b527-91a3871c271d" containerName="extract-utilities" Oct 08 00:10:46 crc kubenswrapper[4967]: I1008 00:10:46.667235 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="f174780f-4c64-427f-b527-91a3871c271d" containerName="extract-utilities" Oct 08 00:10:46 crc kubenswrapper[4967]: E1008 00:10:46.667246 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f79e3965-8fdc-48ba-be6c-2866348f2458" containerName="marketplace-operator" Oct 08 00:10:46 crc kubenswrapper[4967]: I1008 00:10:46.667254 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="f79e3965-8fdc-48ba-be6c-2866348f2458" containerName="marketplace-operator" Oct 08 00:10:46 crc kubenswrapper[4967]: E1008 00:10:46.667267 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3001d43-c530-4c5e-a6e6-27346682ba5c" containerName="extract-content" Oct 08 00:10:46 crc kubenswrapper[4967]: I1008 00:10:46.667274 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3001d43-c530-4c5e-a6e6-27346682ba5c" containerName="extract-content" Oct 08 00:10:46 crc kubenswrapper[4967]: E1008 00:10:46.667282 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3001d43-c530-4c5e-a6e6-27346682ba5c" containerName="registry-server" Oct 08 00:10:46 crc kubenswrapper[4967]: I1008 00:10:46.667289 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3001d43-c530-4c5e-a6e6-27346682ba5c" containerName="registry-server" Oct 08 00:10:46 crc kubenswrapper[4967]: E1008 00:10:46.667300 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32d2d299-2970-40e9-9116-74a074c80eb1" containerName="extract-content" Oct 08 00:10:46 crc kubenswrapper[4967]: I1008 00:10:46.667308 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="32d2d299-2970-40e9-9116-74a074c80eb1" containerName="extract-content" Oct 08 00:10:46 crc kubenswrapper[4967]: E1008 00:10:46.667317 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0340c31-098f-4936-bd5e-4ef17c48a9f1" containerName="extract-utilities" Oct 08 00:10:46 crc kubenswrapper[4967]: I1008 00:10:46.667324 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0340c31-098f-4936-bd5e-4ef17c48a9f1" containerName="extract-utilities" Oct 08 00:10:46 crc kubenswrapper[4967]: E1008 00:10:46.667336 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32d2d299-2970-40e9-9116-74a074c80eb1" containerName="registry-server" Oct 08 00:10:46 crc kubenswrapper[4967]: I1008 00:10:46.667343 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="32d2d299-2970-40e9-9116-74a074c80eb1" containerName="registry-server" Oct 08 00:10:46 crc kubenswrapper[4967]: E1008 00:10:46.667355 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0340c31-098f-4936-bd5e-4ef17c48a9f1" containerName="registry-server" Oct 08 00:10:46 crc kubenswrapper[4967]: I1008 00:10:46.667362 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0340c31-098f-4936-bd5e-4ef17c48a9f1" containerName="registry-server" Oct 08 00:10:46 crc kubenswrapper[4967]: I1008 00:10:46.667467 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="f79e3965-8fdc-48ba-be6c-2866348f2458" containerName="marketplace-operator" Oct 08 00:10:46 crc kubenswrapper[4967]: I1008 00:10:46.667480 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3001d43-c530-4c5e-a6e6-27346682ba5c" containerName="registry-server" Oct 08 00:10:46 crc kubenswrapper[4967]: I1008 00:10:46.667495 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="32d2d299-2970-40e9-9116-74a074c80eb1" containerName="registry-server" Oct 08 00:10:46 crc kubenswrapper[4967]: I1008 00:10:46.667506 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="f174780f-4c64-427f-b527-91a3871c271d" containerName="registry-server" Oct 08 00:10:46 crc kubenswrapper[4967]: I1008 00:10:46.667516 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0340c31-098f-4936-bd5e-4ef17c48a9f1" containerName="registry-server" Oct 08 00:10:46 crc kubenswrapper[4967]: I1008 00:10:46.668295 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pjrvj" Oct 08 00:10:46 crc kubenswrapper[4967]: I1008 00:10:46.670319 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 08 00:10:46 crc kubenswrapper[4967]: I1008 00:10:46.675158 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pjrvj"] Oct 08 00:10:46 crc kubenswrapper[4967]: I1008 00:10:46.811457 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8k5m8\" (UniqueName: \"kubernetes.io/projected/09cce4f0-143f-44b5-8343-56409fa4d0a3-kube-api-access-8k5m8\") pod \"redhat-marketplace-pjrvj\" (UID: \"09cce4f0-143f-44b5-8343-56409fa4d0a3\") " pod="openshift-marketplace/redhat-marketplace-pjrvj" Oct 08 00:10:46 crc kubenswrapper[4967]: I1008 00:10:46.811494 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09cce4f0-143f-44b5-8343-56409fa4d0a3-catalog-content\") pod \"redhat-marketplace-pjrvj\" (UID: \"09cce4f0-143f-44b5-8343-56409fa4d0a3\") " pod="openshift-marketplace/redhat-marketplace-pjrvj" Oct 08 00:10:46 crc kubenswrapper[4967]: I1008 00:10:46.811626 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09cce4f0-143f-44b5-8343-56409fa4d0a3-utilities\") pod \"redhat-marketplace-pjrvj\" (UID: \"09cce4f0-143f-44b5-8343-56409fa4d0a3\") " pod="openshift-marketplace/redhat-marketplace-pjrvj" Oct 08 00:10:46 crc kubenswrapper[4967]: I1008 00:10:46.912323 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8k5m8\" (UniqueName: \"kubernetes.io/projected/09cce4f0-143f-44b5-8343-56409fa4d0a3-kube-api-access-8k5m8\") pod \"redhat-marketplace-pjrvj\" (UID: \"09cce4f0-143f-44b5-8343-56409fa4d0a3\") " pod="openshift-marketplace/redhat-marketplace-pjrvj" Oct 08 00:10:46 crc kubenswrapper[4967]: I1008 00:10:46.912392 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09cce4f0-143f-44b5-8343-56409fa4d0a3-catalog-content\") pod \"redhat-marketplace-pjrvj\" (UID: \"09cce4f0-143f-44b5-8343-56409fa4d0a3\") " pod="openshift-marketplace/redhat-marketplace-pjrvj" Oct 08 00:10:46 crc kubenswrapper[4967]: I1008 00:10:46.912447 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09cce4f0-143f-44b5-8343-56409fa4d0a3-utilities\") pod \"redhat-marketplace-pjrvj\" (UID: \"09cce4f0-143f-44b5-8343-56409fa4d0a3\") " pod="openshift-marketplace/redhat-marketplace-pjrvj" Oct 08 00:10:46 crc kubenswrapper[4967]: I1008 00:10:46.913483 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09cce4f0-143f-44b5-8343-56409fa4d0a3-catalog-content\") pod \"redhat-marketplace-pjrvj\" (UID: \"09cce4f0-143f-44b5-8343-56409fa4d0a3\") " pod="openshift-marketplace/redhat-marketplace-pjrvj" Oct 08 00:10:46 crc kubenswrapper[4967]: I1008 00:10:46.913989 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09cce4f0-143f-44b5-8343-56409fa4d0a3-utilities\") pod \"redhat-marketplace-pjrvj\" (UID: \"09cce4f0-143f-44b5-8343-56409fa4d0a3\") " pod="openshift-marketplace/redhat-marketplace-pjrvj" Oct 08 00:10:46 crc kubenswrapper[4967]: I1008 00:10:46.928648 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8k5m8\" (UniqueName: \"kubernetes.io/projected/09cce4f0-143f-44b5-8343-56409fa4d0a3-kube-api-access-8k5m8\") pod \"redhat-marketplace-pjrvj\" (UID: \"09cce4f0-143f-44b5-8343-56409fa4d0a3\") " pod="openshift-marketplace/redhat-marketplace-pjrvj" Oct 08 00:10:46 crc kubenswrapper[4967]: I1008 00:10:46.988101 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pjrvj" Oct 08 00:10:47 crc kubenswrapper[4967]: I1008 00:10:47.269297 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-6gq5j"] Oct 08 00:10:47 crc kubenswrapper[4967]: I1008 00:10:47.270798 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6gq5j" Oct 08 00:10:47 crc kubenswrapper[4967]: I1008 00:10:47.273468 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 08 00:10:47 crc kubenswrapper[4967]: I1008 00:10:47.282050 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6gq5j"] Oct 08 00:10:47 crc kubenswrapper[4967]: I1008 00:10:47.392320 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pjrvj"] Oct 08 00:10:47 crc kubenswrapper[4967]: W1008 00:10:47.399025 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod09cce4f0_143f_44b5_8343_56409fa4d0a3.slice/crio-e56e99111d4ae2c4e8449a95c1f5a52f0b3923e23568ac7460e2f235ede7eab1 WatchSource:0}: Error finding container e56e99111d4ae2c4e8449a95c1f5a52f0b3923e23568ac7460e2f235ede7eab1: Status 404 returned error can't find the container with id e56e99111d4ae2c4e8449a95c1f5a52f0b3923e23568ac7460e2f235ede7eab1 Oct 08 00:10:47 crc kubenswrapper[4967]: I1008 00:10:47.421338 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17ee9c06-c101-414a-ad72-753c071c0cda-utilities\") pod \"redhat-operators-6gq5j\" (UID: \"17ee9c06-c101-414a-ad72-753c071c0cda\") " pod="openshift-marketplace/redhat-operators-6gq5j" Oct 08 00:10:47 crc kubenswrapper[4967]: I1008 00:10:47.421515 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17ee9c06-c101-414a-ad72-753c071c0cda-catalog-content\") pod \"redhat-operators-6gq5j\" (UID: \"17ee9c06-c101-414a-ad72-753c071c0cda\") " pod="openshift-marketplace/redhat-operators-6gq5j" Oct 08 00:10:47 crc kubenswrapper[4967]: I1008 00:10:47.421577 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5xb7c\" (UniqueName: \"kubernetes.io/projected/17ee9c06-c101-414a-ad72-753c071c0cda-kube-api-access-5xb7c\") pod \"redhat-operators-6gq5j\" (UID: \"17ee9c06-c101-414a-ad72-753c071c0cda\") " pod="openshift-marketplace/redhat-operators-6gq5j" Oct 08 00:10:47 crc kubenswrapper[4967]: I1008 00:10:47.522570 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17ee9c06-c101-414a-ad72-753c071c0cda-utilities\") pod \"redhat-operators-6gq5j\" (UID: \"17ee9c06-c101-414a-ad72-753c071c0cda\") " pod="openshift-marketplace/redhat-operators-6gq5j" Oct 08 00:10:47 crc kubenswrapper[4967]: I1008 00:10:47.522650 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17ee9c06-c101-414a-ad72-753c071c0cda-catalog-content\") pod \"redhat-operators-6gq5j\" (UID: \"17ee9c06-c101-414a-ad72-753c071c0cda\") " pod="openshift-marketplace/redhat-operators-6gq5j" Oct 08 00:10:47 crc kubenswrapper[4967]: I1008 00:10:47.522702 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5xb7c\" (UniqueName: \"kubernetes.io/projected/17ee9c06-c101-414a-ad72-753c071c0cda-kube-api-access-5xb7c\") pod \"redhat-operators-6gq5j\" (UID: \"17ee9c06-c101-414a-ad72-753c071c0cda\") " pod="openshift-marketplace/redhat-operators-6gq5j" Oct 08 00:10:47 crc kubenswrapper[4967]: I1008 00:10:47.523080 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17ee9c06-c101-414a-ad72-753c071c0cda-utilities\") pod \"redhat-operators-6gq5j\" (UID: \"17ee9c06-c101-414a-ad72-753c071c0cda\") " pod="openshift-marketplace/redhat-operators-6gq5j" Oct 08 00:10:47 crc kubenswrapper[4967]: I1008 00:10:47.523378 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17ee9c06-c101-414a-ad72-753c071c0cda-catalog-content\") pod \"redhat-operators-6gq5j\" (UID: \"17ee9c06-c101-414a-ad72-753c071c0cda\") " pod="openshift-marketplace/redhat-operators-6gq5j" Oct 08 00:10:47 crc kubenswrapper[4967]: I1008 00:10:47.544644 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5xb7c\" (UniqueName: \"kubernetes.io/projected/17ee9c06-c101-414a-ad72-753c071c0cda-kube-api-access-5xb7c\") pod \"redhat-operators-6gq5j\" (UID: \"17ee9c06-c101-414a-ad72-753c071c0cda\") " pod="openshift-marketplace/redhat-operators-6gq5j" Oct 08 00:10:47 crc kubenswrapper[4967]: I1008 00:10:47.598003 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6gq5j" Oct 08 00:10:47 crc kubenswrapper[4967]: I1008 00:10:47.800294 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6gq5j"] Oct 08 00:10:47 crc kubenswrapper[4967]: W1008 00:10:47.807136 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod17ee9c06_c101_414a_ad72_753c071c0cda.slice/crio-02765c6eae1c732fd2a97b662fa77109ac198b6a3b6df33523171a597023e874 WatchSource:0}: Error finding container 02765c6eae1c732fd2a97b662fa77109ac198b6a3b6df33523171a597023e874: Status 404 returned error can't find the container with id 02765c6eae1c732fd2a97b662fa77109ac198b6a3b6df33523171a597023e874 Oct 08 00:10:48 crc kubenswrapper[4967]: I1008 00:10:48.311837 4967 generic.go:334] "Generic (PLEG): container finished" podID="17ee9c06-c101-414a-ad72-753c071c0cda" containerID="5e3c7e4327aadf02596cd0b9ac33dc02993d18d1d39e958afdc6b906e8e93e52" exitCode=0 Oct 08 00:10:48 crc kubenswrapper[4967]: I1008 00:10:48.311947 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6gq5j" event={"ID":"17ee9c06-c101-414a-ad72-753c071c0cda","Type":"ContainerDied","Data":"5e3c7e4327aadf02596cd0b9ac33dc02993d18d1d39e958afdc6b906e8e93e52"} Oct 08 00:10:48 crc kubenswrapper[4967]: I1008 00:10:48.312201 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6gq5j" event={"ID":"17ee9c06-c101-414a-ad72-753c071c0cda","Type":"ContainerStarted","Data":"02765c6eae1c732fd2a97b662fa77109ac198b6a3b6df33523171a597023e874"} Oct 08 00:10:48 crc kubenswrapper[4967]: I1008 00:10:48.315000 4967 generic.go:334] "Generic (PLEG): container finished" podID="09cce4f0-143f-44b5-8343-56409fa4d0a3" containerID="8ef952affb669ec11c952345092a78a93b0107f73ef03af4901e9a30007da121" exitCode=0 Oct 08 00:10:48 crc kubenswrapper[4967]: I1008 00:10:48.315044 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pjrvj" event={"ID":"09cce4f0-143f-44b5-8343-56409fa4d0a3","Type":"ContainerDied","Data":"8ef952affb669ec11c952345092a78a93b0107f73ef03af4901e9a30007da121"} Oct 08 00:10:48 crc kubenswrapper[4967]: I1008 00:10:48.315101 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pjrvj" event={"ID":"09cce4f0-143f-44b5-8343-56409fa4d0a3","Type":"ContainerStarted","Data":"e56e99111d4ae2c4e8449a95c1f5a52f0b3923e23568ac7460e2f235ede7eab1"} Oct 08 00:10:49 crc kubenswrapper[4967]: I1008 00:10:49.067627 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-wc5gb"] Oct 08 00:10:49 crc kubenswrapper[4967]: I1008 00:10:49.068854 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wc5gb" Oct 08 00:10:49 crc kubenswrapper[4967]: I1008 00:10:49.071137 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 08 00:10:49 crc kubenswrapper[4967]: I1008 00:10:49.095752 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wc5gb"] Oct 08 00:10:49 crc kubenswrapper[4967]: I1008 00:10:49.153548 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36c8543a-dbe6-4aac-bb98-f39558beb387-catalog-content\") pod \"community-operators-wc5gb\" (UID: \"36c8543a-dbe6-4aac-bb98-f39558beb387\") " pod="openshift-marketplace/community-operators-wc5gb" Oct 08 00:10:49 crc kubenswrapper[4967]: I1008 00:10:49.153600 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36c8543a-dbe6-4aac-bb98-f39558beb387-utilities\") pod \"community-operators-wc5gb\" (UID: \"36c8543a-dbe6-4aac-bb98-f39558beb387\") " pod="openshift-marketplace/community-operators-wc5gb" Oct 08 00:10:49 crc kubenswrapper[4967]: I1008 00:10:49.153636 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wjs4x\" (UniqueName: \"kubernetes.io/projected/36c8543a-dbe6-4aac-bb98-f39558beb387-kube-api-access-wjs4x\") pod \"community-operators-wc5gb\" (UID: \"36c8543a-dbe6-4aac-bb98-f39558beb387\") " pod="openshift-marketplace/community-operators-wc5gb" Oct 08 00:10:49 crc kubenswrapper[4967]: I1008 00:10:49.254482 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wjs4x\" (UniqueName: \"kubernetes.io/projected/36c8543a-dbe6-4aac-bb98-f39558beb387-kube-api-access-wjs4x\") pod \"community-operators-wc5gb\" (UID: \"36c8543a-dbe6-4aac-bb98-f39558beb387\") " pod="openshift-marketplace/community-operators-wc5gb" Oct 08 00:10:49 crc kubenswrapper[4967]: I1008 00:10:49.254654 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36c8543a-dbe6-4aac-bb98-f39558beb387-catalog-content\") pod \"community-operators-wc5gb\" (UID: \"36c8543a-dbe6-4aac-bb98-f39558beb387\") " pod="openshift-marketplace/community-operators-wc5gb" Oct 08 00:10:49 crc kubenswrapper[4967]: I1008 00:10:49.254693 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36c8543a-dbe6-4aac-bb98-f39558beb387-utilities\") pod \"community-operators-wc5gb\" (UID: \"36c8543a-dbe6-4aac-bb98-f39558beb387\") " pod="openshift-marketplace/community-operators-wc5gb" Oct 08 00:10:49 crc kubenswrapper[4967]: I1008 00:10:49.255069 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36c8543a-dbe6-4aac-bb98-f39558beb387-utilities\") pod \"community-operators-wc5gb\" (UID: \"36c8543a-dbe6-4aac-bb98-f39558beb387\") " pod="openshift-marketplace/community-operators-wc5gb" Oct 08 00:10:49 crc kubenswrapper[4967]: I1008 00:10:49.255175 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36c8543a-dbe6-4aac-bb98-f39558beb387-catalog-content\") pod \"community-operators-wc5gb\" (UID: \"36c8543a-dbe6-4aac-bb98-f39558beb387\") " pod="openshift-marketplace/community-operators-wc5gb" Oct 08 00:10:49 crc kubenswrapper[4967]: I1008 00:10:49.285898 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wjs4x\" (UniqueName: \"kubernetes.io/projected/36c8543a-dbe6-4aac-bb98-f39558beb387-kube-api-access-wjs4x\") pod \"community-operators-wc5gb\" (UID: \"36c8543a-dbe6-4aac-bb98-f39558beb387\") " pod="openshift-marketplace/community-operators-wc5gb" Oct 08 00:10:49 crc kubenswrapper[4967]: I1008 00:10:49.410343 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wc5gb" Oct 08 00:10:49 crc kubenswrapper[4967]: I1008 00:10:49.657220 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wc5gb"] Oct 08 00:10:49 crc kubenswrapper[4967]: I1008 00:10:49.672183 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-hk9rd"] Oct 08 00:10:49 crc kubenswrapper[4967]: I1008 00:10:49.673297 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hk9rd" Oct 08 00:10:49 crc kubenswrapper[4967]: I1008 00:10:49.678479 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 08 00:10:49 crc kubenswrapper[4967]: I1008 00:10:49.684653 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hk9rd"] Oct 08 00:10:49 crc kubenswrapper[4967]: W1008 00:10:49.723096 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod36c8543a_dbe6_4aac_bb98_f39558beb387.slice/crio-41de82a0b0e38c23ecda7ff30ab35184c86f7cf887535b553691229b909b3c87 WatchSource:0}: Error finding container 41de82a0b0e38c23ecda7ff30ab35184c86f7cf887535b553691229b909b3c87: Status 404 returned error can't find the container with id 41de82a0b0e38c23ecda7ff30ab35184c86f7cf887535b553691229b909b3c87 Oct 08 00:10:49 crc kubenswrapper[4967]: I1008 00:10:49.861384 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n9rcj\" (UniqueName: \"kubernetes.io/projected/7355878e-878d-494a-98f3-f4a8f6c0c613-kube-api-access-n9rcj\") pod \"certified-operators-hk9rd\" (UID: \"7355878e-878d-494a-98f3-f4a8f6c0c613\") " pod="openshift-marketplace/certified-operators-hk9rd" Oct 08 00:10:49 crc kubenswrapper[4967]: I1008 00:10:49.861481 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7355878e-878d-494a-98f3-f4a8f6c0c613-catalog-content\") pod \"certified-operators-hk9rd\" (UID: \"7355878e-878d-494a-98f3-f4a8f6c0c613\") " pod="openshift-marketplace/certified-operators-hk9rd" Oct 08 00:10:49 crc kubenswrapper[4967]: I1008 00:10:49.861571 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7355878e-878d-494a-98f3-f4a8f6c0c613-utilities\") pod \"certified-operators-hk9rd\" (UID: \"7355878e-878d-494a-98f3-f4a8f6c0c613\") " pod="openshift-marketplace/certified-operators-hk9rd" Oct 08 00:10:49 crc kubenswrapper[4967]: I1008 00:10:49.962327 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7355878e-878d-494a-98f3-f4a8f6c0c613-utilities\") pod \"certified-operators-hk9rd\" (UID: \"7355878e-878d-494a-98f3-f4a8f6c0c613\") " pod="openshift-marketplace/certified-operators-hk9rd" Oct 08 00:10:49 crc kubenswrapper[4967]: I1008 00:10:49.962381 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n9rcj\" (UniqueName: \"kubernetes.io/projected/7355878e-878d-494a-98f3-f4a8f6c0c613-kube-api-access-n9rcj\") pod \"certified-operators-hk9rd\" (UID: \"7355878e-878d-494a-98f3-f4a8f6c0c613\") " pod="openshift-marketplace/certified-operators-hk9rd" Oct 08 00:10:49 crc kubenswrapper[4967]: I1008 00:10:49.962412 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7355878e-878d-494a-98f3-f4a8f6c0c613-catalog-content\") pod \"certified-operators-hk9rd\" (UID: \"7355878e-878d-494a-98f3-f4a8f6c0c613\") " pod="openshift-marketplace/certified-operators-hk9rd" Oct 08 00:10:49 crc kubenswrapper[4967]: I1008 00:10:49.963050 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7355878e-878d-494a-98f3-f4a8f6c0c613-utilities\") pod \"certified-operators-hk9rd\" (UID: \"7355878e-878d-494a-98f3-f4a8f6c0c613\") " pod="openshift-marketplace/certified-operators-hk9rd" Oct 08 00:10:49 crc kubenswrapper[4967]: I1008 00:10:49.963218 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7355878e-878d-494a-98f3-f4a8f6c0c613-catalog-content\") pod \"certified-operators-hk9rd\" (UID: \"7355878e-878d-494a-98f3-f4a8f6c0c613\") " pod="openshift-marketplace/certified-operators-hk9rd" Oct 08 00:10:49 crc kubenswrapper[4967]: I1008 00:10:49.978882 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n9rcj\" (UniqueName: \"kubernetes.io/projected/7355878e-878d-494a-98f3-f4a8f6c0c613-kube-api-access-n9rcj\") pod \"certified-operators-hk9rd\" (UID: \"7355878e-878d-494a-98f3-f4a8f6c0c613\") " pod="openshift-marketplace/certified-operators-hk9rd" Oct 08 00:10:50 crc kubenswrapper[4967]: I1008 00:10:50.031295 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hk9rd" Oct 08 00:10:50 crc kubenswrapper[4967]: I1008 00:10:50.220465 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hk9rd"] Oct 08 00:10:50 crc kubenswrapper[4967]: W1008 00:10:50.227498 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7355878e_878d_494a_98f3_f4a8f6c0c613.slice/crio-08c54f42210904a97ca344922f18ca5b2ef51446fc730a4a3e5bb2e47a102ec7 WatchSource:0}: Error finding container 08c54f42210904a97ca344922f18ca5b2ef51446fc730a4a3e5bb2e47a102ec7: Status 404 returned error can't find the container with id 08c54f42210904a97ca344922f18ca5b2ef51446fc730a4a3e5bb2e47a102ec7 Oct 08 00:10:50 crc kubenswrapper[4967]: I1008 00:10:50.328044 4967 generic.go:334] "Generic (PLEG): container finished" podID="17ee9c06-c101-414a-ad72-753c071c0cda" containerID="43c5a697be910839f82668ccbc01a059a422b104e89dd8652bb18b1230784dda" exitCode=0 Oct 08 00:10:50 crc kubenswrapper[4967]: I1008 00:10:50.328127 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6gq5j" event={"ID":"17ee9c06-c101-414a-ad72-753c071c0cda","Type":"ContainerDied","Data":"43c5a697be910839f82668ccbc01a059a422b104e89dd8652bb18b1230784dda"} Oct 08 00:10:50 crc kubenswrapper[4967]: I1008 00:10:50.329931 4967 generic.go:334] "Generic (PLEG): container finished" podID="09cce4f0-143f-44b5-8343-56409fa4d0a3" containerID="2cff2052c829abc3002ad4560c04638e0e16ee835fea82c4782b0ebbdc8e99b6" exitCode=0 Oct 08 00:10:50 crc kubenswrapper[4967]: I1008 00:10:50.329981 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pjrvj" event={"ID":"09cce4f0-143f-44b5-8343-56409fa4d0a3","Type":"ContainerDied","Data":"2cff2052c829abc3002ad4560c04638e0e16ee835fea82c4782b0ebbdc8e99b6"} Oct 08 00:10:50 crc kubenswrapper[4967]: I1008 00:10:50.331826 4967 generic.go:334] "Generic (PLEG): container finished" podID="36c8543a-dbe6-4aac-bb98-f39558beb387" containerID="8b540946e5e3b3e00134437daa949773ac41db5d92798e2c146eb751fe3c32ff" exitCode=0 Oct 08 00:10:50 crc kubenswrapper[4967]: I1008 00:10:50.331897 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wc5gb" event={"ID":"36c8543a-dbe6-4aac-bb98-f39558beb387","Type":"ContainerDied","Data":"8b540946e5e3b3e00134437daa949773ac41db5d92798e2c146eb751fe3c32ff"} Oct 08 00:10:50 crc kubenswrapper[4967]: I1008 00:10:50.331919 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wc5gb" event={"ID":"36c8543a-dbe6-4aac-bb98-f39558beb387","Type":"ContainerStarted","Data":"41de82a0b0e38c23ecda7ff30ab35184c86f7cf887535b553691229b909b3c87"} Oct 08 00:10:50 crc kubenswrapper[4967]: I1008 00:10:50.339343 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hk9rd" event={"ID":"7355878e-878d-494a-98f3-f4a8f6c0c613","Type":"ContainerStarted","Data":"08c54f42210904a97ca344922f18ca5b2ef51446fc730a4a3e5bb2e47a102ec7"} Oct 08 00:10:51 crc kubenswrapper[4967]: I1008 00:10:51.345739 4967 generic.go:334] "Generic (PLEG): container finished" podID="7355878e-878d-494a-98f3-f4a8f6c0c613" containerID="9be771704492a3c254af5658a099d27b11476044bd8067ab78dd4513e5a0085e" exitCode=0 Oct 08 00:10:51 crc kubenswrapper[4967]: I1008 00:10:51.345782 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hk9rd" event={"ID":"7355878e-878d-494a-98f3-f4a8f6c0c613","Type":"ContainerDied","Data":"9be771704492a3c254af5658a099d27b11476044bd8067ab78dd4513e5a0085e"} Oct 08 00:10:51 crc kubenswrapper[4967]: I1008 00:10:51.350846 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6gq5j" event={"ID":"17ee9c06-c101-414a-ad72-753c071c0cda","Type":"ContainerStarted","Data":"80ade5e84c70f3d0df59557c080587b862fce917a70e7c10bc954b8ff9faa43a"} Oct 08 00:10:51 crc kubenswrapper[4967]: I1008 00:10:51.377436 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-6gq5j" podStartSLOduration=1.917327899 podStartE2EDuration="4.377416625s" podCreationTimestamp="2025-10-08 00:10:47 +0000 UTC" firstStartedPulling="2025-10-08 00:10:48.313382183 +0000 UTC m=+246.963916832" lastFinishedPulling="2025-10-08 00:10:50.773470909 +0000 UTC m=+249.424005558" observedRunningTime="2025-10-08 00:10:51.376551626 +0000 UTC m=+250.027086275" watchObservedRunningTime="2025-10-08 00:10:51.377416625 +0000 UTC m=+250.027951274" Oct 08 00:10:52 crc kubenswrapper[4967]: I1008 00:10:52.359509 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pjrvj" event={"ID":"09cce4f0-143f-44b5-8343-56409fa4d0a3","Type":"ContainerStarted","Data":"a474d3f13903ad7e8d8340d2c559e1b6d672592b2d1e0aaf231f717dfafb3e9b"} Oct 08 00:10:52 crc kubenswrapper[4967]: I1008 00:10:52.362839 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wc5gb" event={"ID":"36c8543a-dbe6-4aac-bb98-f39558beb387","Type":"ContainerStarted","Data":"c75c84a6645055d6ad00cae5407ab40d7bb66138896bd2b4531bcd84f689155b"} Oct 08 00:10:52 crc kubenswrapper[4967]: I1008 00:10:52.378648 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-pjrvj" podStartSLOduration=3.694194087 podStartE2EDuration="6.378630367s" podCreationTimestamp="2025-10-08 00:10:46 +0000 UTC" firstStartedPulling="2025-10-08 00:10:48.316134666 +0000 UTC m=+246.966669345" lastFinishedPulling="2025-10-08 00:10:51.000570916 +0000 UTC m=+249.651105625" observedRunningTime="2025-10-08 00:10:52.375126277 +0000 UTC m=+251.025660926" watchObservedRunningTime="2025-10-08 00:10:52.378630367 +0000 UTC m=+251.029165016" Oct 08 00:10:53 crc kubenswrapper[4967]: I1008 00:10:53.369700 4967 generic.go:334] "Generic (PLEG): container finished" podID="36c8543a-dbe6-4aac-bb98-f39558beb387" containerID="c75c84a6645055d6ad00cae5407ab40d7bb66138896bd2b4531bcd84f689155b" exitCode=0 Oct 08 00:10:53 crc kubenswrapper[4967]: I1008 00:10:53.369829 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wc5gb" event={"ID":"36c8543a-dbe6-4aac-bb98-f39558beb387","Type":"ContainerDied","Data":"c75c84a6645055d6ad00cae5407ab40d7bb66138896bd2b4531bcd84f689155b"} Oct 08 00:10:56 crc kubenswrapper[4967]: I1008 00:10:56.385954 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wc5gb" event={"ID":"36c8543a-dbe6-4aac-bb98-f39558beb387","Type":"ContainerStarted","Data":"364aa85a8040c4409dbafa8f7a1f0c7d0d680420a1d64bd2e8d07c637f08946c"} Oct 08 00:10:56 crc kubenswrapper[4967]: I1008 00:10:56.389803 4967 generic.go:334] "Generic (PLEG): container finished" podID="7355878e-878d-494a-98f3-f4a8f6c0c613" containerID="52809ee0e8cbe0f068e962bbe61d0af83bc57e86c389b26aea049e47a8a62750" exitCode=0 Oct 08 00:10:56 crc kubenswrapper[4967]: I1008 00:10:56.389839 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hk9rd" event={"ID":"7355878e-878d-494a-98f3-f4a8f6c0c613","Type":"ContainerDied","Data":"52809ee0e8cbe0f068e962bbe61d0af83bc57e86c389b26aea049e47a8a62750"} Oct 08 00:10:56 crc kubenswrapper[4967]: I1008 00:10:56.404397 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-wc5gb" podStartSLOduration=2.150316358 podStartE2EDuration="7.404378796s" podCreationTimestamp="2025-10-08 00:10:49 +0000 UTC" firstStartedPulling="2025-10-08 00:10:50.332867867 +0000 UTC m=+248.983402516" lastFinishedPulling="2025-10-08 00:10:55.586930305 +0000 UTC m=+254.237464954" observedRunningTime="2025-10-08 00:10:56.401891822 +0000 UTC m=+255.052426491" watchObservedRunningTime="2025-10-08 00:10:56.404378796 +0000 UTC m=+255.054913445" Oct 08 00:10:56 crc kubenswrapper[4967]: I1008 00:10:56.988795 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-pjrvj" Oct 08 00:10:56 crc kubenswrapper[4967]: I1008 00:10:56.988930 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-pjrvj" Oct 08 00:10:57 crc kubenswrapper[4967]: I1008 00:10:57.026556 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-pjrvj" Oct 08 00:10:57 crc kubenswrapper[4967]: I1008 00:10:57.406186 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hk9rd" event={"ID":"7355878e-878d-494a-98f3-f4a8f6c0c613","Type":"ContainerStarted","Data":"b5122b9952fea05fb96759e93e4cbfcdf396735cc1b600f9c04736a9735fc67f"} Oct 08 00:10:57 crc kubenswrapper[4967]: I1008 00:10:57.456738 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-pjrvj" Oct 08 00:10:57 crc kubenswrapper[4967]: I1008 00:10:57.476563 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-hk9rd" podStartSLOduration=2.969074803 podStartE2EDuration="8.476545485s" podCreationTimestamp="2025-10-08 00:10:49 +0000 UTC" firstStartedPulling="2025-10-08 00:10:51.348231231 +0000 UTC m=+249.998765880" lastFinishedPulling="2025-10-08 00:10:56.855701893 +0000 UTC m=+255.506236562" observedRunningTime="2025-10-08 00:10:57.423801739 +0000 UTC m=+256.074336388" watchObservedRunningTime="2025-10-08 00:10:57.476545485 +0000 UTC m=+256.127080124" Oct 08 00:10:57 crc kubenswrapper[4967]: I1008 00:10:57.598910 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-6gq5j" Oct 08 00:10:57 crc kubenswrapper[4967]: I1008 00:10:57.598964 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-6gq5j" Oct 08 00:10:57 crc kubenswrapper[4967]: I1008 00:10:57.647109 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-6gq5j" Oct 08 00:10:58 crc kubenswrapper[4967]: I1008 00:10:58.442753 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-6gq5j" Oct 08 00:10:59 crc kubenswrapper[4967]: I1008 00:10:59.410604 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-wc5gb" Oct 08 00:10:59 crc kubenswrapper[4967]: I1008 00:10:59.411077 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-wc5gb" Oct 08 00:10:59 crc kubenswrapper[4967]: I1008 00:10:59.475365 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-wc5gb" Oct 08 00:11:00 crc kubenswrapper[4967]: I1008 00:11:00.032071 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-hk9rd" Oct 08 00:11:00 crc kubenswrapper[4967]: I1008 00:11:00.032115 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-hk9rd" Oct 08 00:11:00 crc kubenswrapper[4967]: I1008 00:11:00.068234 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-hk9rd" Oct 08 00:11:00 crc kubenswrapper[4967]: I1008 00:11:00.467652 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-wc5gb" Oct 08 00:11:10 crc kubenswrapper[4967]: I1008 00:11:10.077148 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-hk9rd" Oct 08 00:12:36 crc kubenswrapper[4967]: I1008 00:12:36.920094 4967 patch_prober.go:28] interesting pod/machine-config-daemon-xvdtp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 00:12:36 crc kubenswrapper[4967]: I1008 00:12:36.921139 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" podUID="9461162a-c27d-465e-9ca7-2d3936877cd8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 00:13:06 crc kubenswrapper[4967]: I1008 00:13:06.920892 4967 patch_prober.go:28] interesting pod/machine-config-daemon-xvdtp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 00:13:06 crc kubenswrapper[4967]: I1008 00:13:06.921824 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" podUID="9461162a-c27d-465e-9ca7-2d3936877cd8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 00:13:36 crc kubenswrapper[4967]: I1008 00:13:36.920082 4967 patch_prober.go:28] interesting pod/machine-config-daemon-xvdtp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 00:13:36 crc kubenswrapper[4967]: I1008 00:13:36.920719 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" podUID="9461162a-c27d-465e-9ca7-2d3936877cd8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 00:13:36 crc kubenswrapper[4967]: I1008 00:13:36.920780 4967 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" Oct 08 00:13:36 crc kubenswrapper[4967]: I1008 00:13:36.921541 4967 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d566016b7da391d83d191f50824299b48d2b961375cdeea4e96f5ac21ff9991f"} pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 00:13:36 crc kubenswrapper[4967]: I1008 00:13:36.921637 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" podUID="9461162a-c27d-465e-9ca7-2d3936877cd8" containerName="machine-config-daemon" containerID="cri-o://d566016b7da391d83d191f50824299b48d2b961375cdeea4e96f5ac21ff9991f" gracePeriod=600 Oct 08 00:13:37 crc kubenswrapper[4967]: I1008 00:13:37.460605 4967 generic.go:334] "Generic (PLEG): container finished" podID="9461162a-c27d-465e-9ca7-2d3936877cd8" containerID="d566016b7da391d83d191f50824299b48d2b961375cdeea4e96f5ac21ff9991f" exitCode=0 Oct 08 00:13:37 crc kubenswrapper[4967]: I1008 00:13:37.460715 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" event={"ID":"9461162a-c27d-465e-9ca7-2d3936877cd8","Type":"ContainerDied","Data":"d566016b7da391d83d191f50824299b48d2b961375cdeea4e96f5ac21ff9991f"} Oct 08 00:13:37 crc kubenswrapper[4967]: I1008 00:13:37.461385 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" event={"ID":"9461162a-c27d-465e-9ca7-2d3936877cd8","Type":"ContainerStarted","Data":"f9c94f180083f2a4c120cfed331389932c804613b3fbe7d131d60b469072cb08"} Oct 08 00:13:37 crc kubenswrapper[4967]: I1008 00:13:37.461506 4967 scope.go:117] "RemoveContainer" containerID="28760af04ef321d0f91b2bce3ad074d44b63dabf676cc37f3be6aa8726341a2f" Oct 08 00:14:10 crc kubenswrapper[4967]: I1008 00:14:10.952826 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-4fs9r"] Oct 08 00:14:10 crc kubenswrapper[4967]: I1008 00:14:10.954971 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-4fs9r" Oct 08 00:14:10 crc kubenswrapper[4967]: I1008 00:14:10.967318 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-4fs9r"] Oct 08 00:14:11 crc kubenswrapper[4967]: I1008 00:14:11.071272 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n8xm7\" (UniqueName: \"kubernetes.io/projected/866925ec-2df6-4dcc-93d9-b239a8155563-kube-api-access-n8xm7\") pod \"image-registry-66df7c8f76-4fs9r\" (UID: \"866925ec-2df6-4dcc-93d9-b239a8155563\") " pod="openshift-image-registry/image-registry-66df7c8f76-4fs9r" Oct 08 00:14:11 crc kubenswrapper[4967]: I1008 00:14:11.071538 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/866925ec-2df6-4dcc-93d9-b239a8155563-registry-tls\") pod \"image-registry-66df7c8f76-4fs9r\" (UID: \"866925ec-2df6-4dcc-93d9-b239a8155563\") " pod="openshift-image-registry/image-registry-66df7c8f76-4fs9r" Oct 08 00:14:11 crc kubenswrapper[4967]: I1008 00:14:11.071568 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/866925ec-2df6-4dcc-93d9-b239a8155563-registry-certificates\") pod \"image-registry-66df7c8f76-4fs9r\" (UID: \"866925ec-2df6-4dcc-93d9-b239a8155563\") " pod="openshift-image-registry/image-registry-66df7c8f76-4fs9r" Oct 08 00:14:11 crc kubenswrapper[4967]: I1008 00:14:11.071622 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-4fs9r\" (UID: \"866925ec-2df6-4dcc-93d9-b239a8155563\") " pod="openshift-image-registry/image-registry-66df7c8f76-4fs9r" Oct 08 00:14:11 crc kubenswrapper[4967]: I1008 00:14:11.071650 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/866925ec-2df6-4dcc-93d9-b239a8155563-bound-sa-token\") pod \"image-registry-66df7c8f76-4fs9r\" (UID: \"866925ec-2df6-4dcc-93d9-b239a8155563\") " pod="openshift-image-registry/image-registry-66df7c8f76-4fs9r" Oct 08 00:14:11 crc kubenswrapper[4967]: I1008 00:14:11.071678 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/866925ec-2df6-4dcc-93d9-b239a8155563-trusted-ca\") pod \"image-registry-66df7c8f76-4fs9r\" (UID: \"866925ec-2df6-4dcc-93d9-b239a8155563\") " pod="openshift-image-registry/image-registry-66df7c8f76-4fs9r" Oct 08 00:14:11 crc kubenswrapper[4967]: I1008 00:14:11.071699 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/866925ec-2df6-4dcc-93d9-b239a8155563-installation-pull-secrets\") pod \"image-registry-66df7c8f76-4fs9r\" (UID: \"866925ec-2df6-4dcc-93d9-b239a8155563\") " pod="openshift-image-registry/image-registry-66df7c8f76-4fs9r" Oct 08 00:14:11 crc kubenswrapper[4967]: I1008 00:14:11.071772 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/866925ec-2df6-4dcc-93d9-b239a8155563-ca-trust-extracted\") pod \"image-registry-66df7c8f76-4fs9r\" (UID: \"866925ec-2df6-4dcc-93d9-b239a8155563\") " pod="openshift-image-registry/image-registry-66df7c8f76-4fs9r" Oct 08 00:14:11 crc kubenswrapper[4967]: I1008 00:14:11.114609 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-4fs9r\" (UID: \"866925ec-2df6-4dcc-93d9-b239a8155563\") " pod="openshift-image-registry/image-registry-66df7c8f76-4fs9r" Oct 08 00:14:11 crc kubenswrapper[4967]: I1008 00:14:11.172868 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/866925ec-2df6-4dcc-93d9-b239a8155563-registry-certificates\") pod \"image-registry-66df7c8f76-4fs9r\" (UID: \"866925ec-2df6-4dcc-93d9-b239a8155563\") " pod="openshift-image-registry/image-registry-66df7c8f76-4fs9r" Oct 08 00:14:11 crc kubenswrapper[4967]: I1008 00:14:11.172968 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/866925ec-2df6-4dcc-93d9-b239a8155563-bound-sa-token\") pod \"image-registry-66df7c8f76-4fs9r\" (UID: \"866925ec-2df6-4dcc-93d9-b239a8155563\") " pod="openshift-image-registry/image-registry-66df7c8f76-4fs9r" Oct 08 00:14:11 crc kubenswrapper[4967]: I1008 00:14:11.172997 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/866925ec-2df6-4dcc-93d9-b239a8155563-trusted-ca\") pod \"image-registry-66df7c8f76-4fs9r\" (UID: \"866925ec-2df6-4dcc-93d9-b239a8155563\") " pod="openshift-image-registry/image-registry-66df7c8f76-4fs9r" Oct 08 00:14:11 crc kubenswrapper[4967]: I1008 00:14:11.173019 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/866925ec-2df6-4dcc-93d9-b239a8155563-installation-pull-secrets\") pod \"image-registry-66df7c8f76-4fs9r\" (UID: \"866925ec-2df6-4dcc-93d9-b239a8155563\") " pod="openshift-image-registry/image-registry-66df7c8f76-4fs9r" Oct 08 00:14:11 crc kubenswrapper[4967]: I1008 00:14:11.173043 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/866925ec-2df6-4dcc-93d9-b239a8155563-ca-trust-extracted\") pod \"image-registry-66df7c8f76-4fs9r\" (UID: \"866925ec-2df6-4dcc-93d9-b239a8155563\") " pod="openshift-image-registry/image-registry-66df7c8f76-4fs9r" Oct 08 00:14:11 crc kubenswrapper[4967]: I1008 00:14:11.173079 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n8xm7\" (UniqueName: \"kubernetes.io/projected/866925ec-2df6-4dcc-93d9-b239a8155563-kube-api-access-n8xm7\") pod \"image-registry-66df7c8f76-4fs9r\" (UID: \"866925ec-2df6-4dcc-93d9-b239a8155563\") " pod="openshift-image-registry/image-registry-66df7c8f76-4fs9r" Oct 08 00:14:11 crc kubenswrapper[4967]: I1008 00:14:11.173104 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/866925ec-2df6-4dcc-93d9-b239a8155563-registry-tls\") pod \"image-registry-66df7c8f76-4fs9r\" (UID: \"866925ec-2df6-4dcc-93d9-b239a8155563\") " pod="openshift-image-registry/image-registry-66df7c8f76-4fs9r" Oct 08 00:14:11 crc kubenswrapper[4967]: I1008 00:14:11.174716 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/866925ec-2df6-4dcc-93d9-b239a8155563-ca-trust-extracted\") pod \"image-registry-66df7c8f76-4fs9r\" (UID: \"866925ec-2df6-4dcc-93d9-b239a8155563\") " pod="openshift-image-registry/image-registry-66df7c8f76-4fs9r" Oct 08 00:14:11 crc kubenswrapper[4967]: I1008 00:14:11.175917 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/866925ec-2df6-4dcc-93d9-b239a8155563-registry-certificates\") pod \"image-registry-66df7c8f76-4fs9r\" (UID: \"866925ec-2df6-4dcc-93d9-b239a8155563\") " pod="openshift-image-registry/image-registry-66df7c8f76-4fs9r" Oct 08 00:14:11 crc kubenswrapper[4967]: I1008 00:14:11.176372 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/866925ec-2df6-4dcc-93d9-b239a8155563-trusted-ca\") pod \"image-registry-66df7c8f76-4fs9r\" (UID: \"866925ec-2df6-4dcc-93d9-b239a8155563\") " pod="openshift-image-registry/image-registry-66df7c8f76-4fs9r" Oct 08 00:14:11 crc kubenswrapper[4967]: I1008 00:14:11.180529 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/866925ec-2df6-4dcc-93d9-b239a8155563-installation-pull-secrets\") pod \"image-registry-66df7c8f76-4fs9r\" (UID: \"866925ec-2df6-4dcc-93d9-b239a8155563\") " pod="openshift-image-registry/image-registry-66df7c8f76-4fs9r" Oct 08 00:14:11 crc kubenswrapper[4967]: I1008 00:14:11.180605 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/866925ec-2df6-4dcc-93d9-b239a8155563-registry-tls\") pod \"image-registry-66df7c8f76-4fs9r\" (UID: \"866925ec-2df6-4dcc-93d9-b239a8155563\") " pod="openshift-image-registry/image-registry-66df7c8f76-4fs9r" Oct 08 00:14:11 crc kubenswrapper[4967]: I1008 00:14:11.188724 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n8xm7\" (UniqueName: \"kubernetes.io/projected/866925ec-2df6-4dcc-93d9-b239a8155563-kube-api-access-n8xm7\") pod \"image-registry-66df7c8f76-4fs9r\" (UID: \"866925ec-2df6-4dcc-93d9-b239a8155563\") " pod="openshift-image-registry/image-registry-66df7c8f76-4fs9r" Oct 08 00:14:11 crc kubenswrapper[4967]: I1008 00:14:11.193421 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/866925ec-2df6-4dcc-93d9-b239a8155563-bound-sa-token\") pod \"image-registry-66df7c8f76-4fs9r\" (UID: \"866925ec-2df6-4dcc-93d9-b239a8155563\") " pod="openshift-image-registry/image-registry-66df7c8f76-4fs9r" Oct 08 00:14:11 crc kubenswrapper[4967]: I1008 00:14:11.274712 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-4fs9r" Oct 08 00:14:11 crc kubenswrapper[4967]: I1008 00:14:11.540242 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-4fs9r"] Oct 08 00:14:11 crc kubenswrapper[4967]: I1008 00:14:11.691592 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-4fs9r" event={"ID":"866925ec-2df6-4dcc-93d9-b239a8155563","Type":"ContainerStarted","Data":"916350171b307324eb8623db7761a582d5042ab42f7eb88e50571e0b3de8f728"} Oct 08 00:14:12 crc kubenswrapper[4967]: I1008 00:14:12.703003 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-4fs9r" event={"ID":"866925ec-2df6-4dcc-93d9-b239a8155563","Type":"ContainerStarted","Data":"9a8f1211dfe221e9660d1b52a4a63c2b9f3874ac0ebe35dbd01185d45e99182d"} Oct 08 00:14:12 crc kubenswrapper[4967]: I1008 00:14:12.703218 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-4fs9r" Oct 08 00:14:12 crc kubenswrapper[4967]: I1008 00:14:12.737405 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-4fs9r" podStartSLOduration=2.73737598 podStartE2EDuration="2.73737598s" podCreationTimestamp="2025-10-08 00:14:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:14:12.734593801 +0000 UTC m=+451.385128480" watchObservedRunningTime="2025-10-08 00:14:12.73737598 +0000 UTC m=+451.387910659" Oct 08 00:14:31 crc kubenswrapper[4967]: I1008 00:14:31.287568 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-4fs9r" Oct 08 00:14:31 crc kubenswrapper[4967]: I1008 00:14:31.370934 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-8d8x5"] Oct 08 00:14:56 crc kubenswrapper[4967]: I1008 00:14:56.436963 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" podUID="1f06a5d1-8ff8-4d73-8939-0fac4252ea9c" containerName="registry" containerID="cri-o://b8d42b81a779b7bfaa854fb91cbc2f3948dc9bc6aac2b6c0fd684e856fc1871e" gracePeriod=30 Oct 08 00:14:56 crc kubenswrapper[4967]: I1008 00:14:56.895372 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:14:57 crc kubenswrapper[4967]: I1008 00:14:57.001946 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/1f06a5d1-8ff8-4d73-8939-0fac4252ea9c-ca-trust-extracted\") pod \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " Oct 08 00:14:57 crc kubenswrapper[4967]: I1008 00:14:57.002254 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " Oct 08 00:14:57 crc kubenswrapper[4967]: I1008 00:14:57.002348 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1f06a5d1-8ff8-4d73-8939-0fac4252ea9c-bound-sa-token\") pod \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " Oct 08 00:14:57 crc kubenswrapper[4967]: I1008 00:14:57.002398 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/1f06a5d1-8ff8-4d73-8939-0fac4252ea9c-registry-certificates\") pod \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " Oct 08 00:14:57 crc kubenswrapper[4967]: I1008 00:14:57.002515 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/1f06a5d1-8ff8-4d73-8939-0fac4252ea9c-installation-pull-secrets\") pod \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " Oct 08 00:14:57 crc kubenswrapper[4967]: I1008 00:14:57.002644 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b6lf6\" (UniqueName: \"kubernetes.io/projected/1f06a5d1-8ff8-4d73-8939-0fac4252ea9c-kube-api-access-b6lf6\") pod \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " Oct 08 00:14:57 crc kubenswrapper[4967]: I1008 00:14:57.002846 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1f06a5d1-8ff8-4d73-8939-0fac4252ea9c-trusted-ca\") pod \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " Oct 08 00:14:57 crc kubenswrapper[4967]: I1008 00:14:57.003064 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/1f06a5d1-8ff8-4d73-8939-0fac4252ea9c-registry-tls\") pod \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\" (UID: \"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c\") " Oct 08 00:14:57 crc kubenswrapper[4967]: I1008 00:14:57.003746 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1f06a5d1-8ff8-4d73-8939-0fac4252ea9c-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "1f06a5d1-8ff8-4d73-8939-0fac4252ea9c" (UID: "1f06a5d1-8ff8-4d73-8939-0fac4252ea9c"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:14:57 crc kubenswrapper[4967]: I1008 00:14:57.003780 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1f06a5d1-8ff8-4d73-8939-0fac4252ea9c-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "1f06a5d1-8ff8-4d73-8939-0fac4252ea9c" (UID: "1f06a5d1-8ff8-4d73-8939-0fac4252ea9c"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:14:57 crc kubenswrapper[4967]: I1008 00:14:57.009296 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f06a5d1-8ff8-4d73-8939-0fac4252ea9c-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "1f06a5d1-8ff8-4d73-8939-0fac4252ea9c" (UID: "1f06a5d1-8ff8-4d73-8939-0fac4252ea9c"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:14:57 crc kubenswrapper[4967]: I1008 00:14:57.012454 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f06a5d1-8ff8-4d73-8939-0fac4252ea9c-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "1f06a5d1-8ff8-4d73-8939-0fac4252ea9c" (UID: "1f06a5d1-8ff8-4d73-8939-0fac4252ea9c"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:14:57 crc kubenswrapper[4967]: I1008 00:14:57.015928 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f06a5d1-8ff8-4d73-8939-0fac4252ea9c-kube-api-access-b6lf6" (OuterVolumeSpecName: "kube-api-access-b6lf6") pod "1f06a5d1-8ff8-4d73-8939-0fac4252ea9c" (UID: "1f06a5d1-8ff8-4d73-8939-0fac4252ea9c"). InnerVolumeSpecName "kube-api-access-b6lf6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:14:57 crc kubenswrapper[4967]: I1008 00:14:57.016962 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f06a5d1-8ff8-4d73-8939-0fac4252ea9c-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "1f06a5d1-8ff8-4d73-8939-0fac4252ea9c" (UID: "1f06a5d1-8ff8-4d73-8939-0fac4252ea9c"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:14:57 crc kubenswrapper[4967]: I1008 00:14:57.019344 4967 generic.go:334] "Generic (PLEG): container finished" podID="1f06a5d1-8ff8-4d73-8939-0fac4252ea9c" containerID="b8d42b81a779b7bfaa854fb91cbc2f3948dc9bc6aac2b6c0fd684e856fc1871e" exitCode=0 Oct 08 00:14:57 crc kubenswrapper[4967]: I1008 00:14:57.019435 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" event={"ID":"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c","Type":"ContainerDied","Data":"b8d42b81a779b7bfaa854fb91cbc2f3948dc9bc6aac2b6c0fd684e856fc1871e"} Oct 08 00:14:57 crc kubenswrapper[4967]: I1008 00:14:57.019444 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" Oct 08 00:14:57 crc kubenswrapper[4967]: I1008 00:14:57.019490 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-8d8x5" event={"ID":"1f06a5d1-8ff8-4d73-8939-0fac4252ea9c","Type":"ContainerDied","Data":"2df3f6aadad5e3c63cc4ae1da8849d34672ca81be52b38a43f9cd0034ab5018c"} Oct 08 00:14:57 crc kubenswrapper[4967]: I1008 00:14:57.019528 4967 scope.go:117] "RemoveContainer" containerID="b8d42b81a779b7bfaa854fb91cbc2f3948dc9bc6aac2b6c0fd684e856fc1871e" Oct 08 00:14:57 crc kubenswrapper[4967]: I1008 00:14:57.024511 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "1f06a5d1-8ff8-4d73-8939-0fac4252ea9c" (UID: "1f06a5d1-8ff8-4d73-8939-0fac4252ea9c"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 08 00:14:57 crc kubenswrapper[4967]: I1008 00:14:57.030433 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1f06a5d1-8ff8-4d73-8939-0fac4252ea9c-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "1f06a5d1-8ff8-4d73-8939-0fac4252ea9c" (UID: "1f06a5d1-8ff8-4d73-8939-0fac4252ea9c"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:14:57 crc kubenswrapper[4967]: I1008 00:14:57.085735 4967 scope.go:117] "RemoveContainer" containerID="b8d42b81a779b7bfaa854fb91cbc2f3948dc9bc6aac2b6c0fd684e856fc1871e" Oct 08 00:14:57 crc kubenswrapper[4967]: E1008 00:14:57.086311 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b8d42b81a779b7bfaa854fb91cbc2f3948dc9bc6aac2b6c0fd684e856fc1871e\": container with ID starting with b8d42b81a779b7bfaa854fb91cbc2f3948dc9bc6aac2b6c0fd684e856fc1871e not found: ID does not exist" containerID="b8d42b81a779b7bfaa854fb91cbc2f3948dc9bc6aac2b6c0fd684e856fc1871e" Oct 08 00:14:57 crc kubenswrapper[4967]: I1008 00:14:57.086348 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b8d42b81a779b7bfaa854fb91cbc2f3948dc9bc6aac2b6c0fd684e856fc1871e"} err="failed to get container status \"b8d42b81a779b7bfaa854fb91cbc2f3948dc9bc6aac2b6c0fd684e856fc1871e\": rpc error: code = NotFound desc = could not find container \"b8d42b81a779b7bfaa854fb91cbc2f3948dc9bc6aac2b6c0fd684e856fc1871e\": container with ID starting with b8d42b81a779b7bfaa854fb91cbc2f3948dc9bc6aac2b6c0fd684e856fc1871e not found: ID does not exist" Oct 08 00:14:57 crc kubenswrapper[4967]: I1008 00:14:57.105692 4967 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/1f06a5d1-8ff8-4d73-8939-0fac4252ea9c-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 08 00:14:57 crc kubenswrapper[4967]: I1008 00:14:57.105760 4967 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1f06a5d1-8ff8-4d73-8939-0fac4252ea9c-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 08 00:14:57 crc kubenswrapper[4967]: I1008 00:14:57.105790 4967 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/1f06a5d1-8ff8-4d73-8939-0fac4252ea9c-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 08 00:14:57 crc kubenswrapper[4967]: I1008 00:14:57.105825 4967 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/1f06a5d1-8ff8-4d73-8939-0fac4252ea9c-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 08 00:14:57 crc kubenswrapper[4967]: I1008 00:14:57.105843 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b6lf6\" (UniqueName: \"kubernetes.io/projected/1f06a5d1-8ff8-4d73-8939-0fac4252ea9c-kube-api-access-b6lf6\") on node \"crc\" DevicePath \"\"" Oct 08 00:14:57 crc kubenswrapper[4967]: I1008 00:14:57.105862 4967 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1f06a5d1-8ff8-4d73-8939-0fac4252ea9c-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 08 00:14:57 crc kubenswrapper[4967]: I1008 00:14:57.105879 4967 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/1f06a5d1-8ff8-4d73-8939-0fac4252ea9c-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 08 00:14:57 crc kubenswrapper[4967]: I1008 00:14:57.359430 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-8d8x5"] Oct 08 00:14:57 crc kubenswrapper[4967]: I1008 00:14:57.363785 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-8d8x5"] Oct 08 00:14:58 crc kubenswrapper[4967]: I1008 00:14:58.127869 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1f06a5d1-8ff8-4d73-8939-0fac4252ea9c" path="/var/lib/kubelet/pods/1f06a5d1-8ff8-4d73-8939-0fac4252ea9c/volumes" Oct 08 00:15:00 crc kubenswrapper[4967]: I1008 00:15:00.156307 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331375-vgl2p"] Oct 08 00:15:00 crc kubenswrapper[4967]: E1008 00:15:00.156833 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f06a5d1-8ff8-4d73-8939-0fac4252ea9c" containerName="registry" Oct 08 00:15:00 crc kubenswrapper[4967]: I1008 00:15:00.156867 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f06a5d1-8ff8-4d73-8939-0fac4252ea9c" containerName="registry" Oct 08 00:15:00 crc kubenswrapper[4967]: I1008 00:15:00.156987 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f06a5d1-8ff8-4d73-8939-0fac4252ea9c" containerName="registry" Oct 08 00:15:00 crc kubenswrapper[4967]: I1008 00:15:00.157434 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331375-vgl2p" Oct 08 00:15:00 crc kubenswrapper[4967]: I1008 00:15:00.160333 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 08 00:15:00 crc kubenswrapper[4967]: I1008 00:15:00.160938 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 08 00:15:00 crc kubenswrapper[4967]: I1008 00:15:00.203307 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331375-vgl2p"] Oct 08 00:15:00 crc kubenswrapper[4967]: I1008 00:15:00.267977 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1f02cac1-b6b3-4a30-b2de-a2eb10a1f359-config-volume\") pod \"collect-profiles-29331375-vgl2p\" (UID: \"1f02cac1-b6b3-4a30-b2de-a2eb10a1f359\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331375-vgl2p" Oct 08 00:15:00 crc kubenswrapper[4967]: I1008 00:15:00.268029 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f4296\" (UniqueName: \"kubernetes.io/projected/1f02cac1-b6b3-4a30-b2de-a2eb10a1f359-kube-api-access-f4296\") pod \"collect-profiles-29331375-vgl2p\" (UID: \"1f02cac1-b6b3-4a30-b2de-a2eb10a1f359\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331375-vgl2p" Oct 08 00:15:00 crc kubenswrapper[4967]: I1008 00:15:00.268050 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1f02cac1-b6b3-4a30-b2de-a2eb10a1f359-secret-volume\") pod \"collect-profiles-29331375-vgl2p\" (UID: \"1f02cac1-b6b3-4a30-b2de-a2eb10a1f359\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331375-vgl2p" Oct 08 00:15:00 crc kubenswrapper[4967]: I1008 00:15:00.369782 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1f02cac1-b6b3-4a30-b2de-a2eb10a1f359-config-volume\") pod \"collect-profiles-29331375-vgl2p\" (UID: \"1f02cac1-b6b3-4a30-b2de-a2eb10a1f359\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331375-vgl2p" Oct 08 00:15:00 crc kubenswrapper[4967]: I1008 00:15:00.369873 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f4296\" (UniqueName: \"kubernetes.io/projected/1f02cac1-b6b3-4a30-b2de-a2eb10a1f359-kube-api-access-f4296\") pod \"collect-profiles-29331375-vgl2p\" (UID: \"1f02cac1-b6b3-4a30-b2de-a2eb10a1f359\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331375-vgl2p" Oct 08 00:15:00 crc kubenswrapper[4967]: I1008 00:15:00.369916 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1f02cac1-b6b3-4a30-b2de-a2eb10a1f359-secret-volume\") pod \"collect-profiles-29331375-vgl2p\" (UID: \"1f02cac1-b6b3-4a30-b2de-a2eb10a1f359\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331375-vgl2p" Oct 08 00:15:00 crc kubenswrapper[4967]: I1008 00:15:00.370755 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1f02cac1-b6b3-4a30-b2de-a2eb10a1f359-config-volume\") pod \"collect-profiles-29331375-vgl2p\" (UID: \"1f02cac1-b6b3-4a30-b2de-a2eb10a1f359\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331375-vgl2p" Oct 08 00:15:00 crc kubenswrapper[4967]: I1008 00:15:00.375110 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1f02cac1-b6b3-4a30-b2de-a2eb10a1f359-secret-volume\") pod \"collect-profiles-29331375-vgl2p\" (UID: \"1f02cac1-b6b3-4a30-b2de-a2eb10a1f359\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331375-vgl2p" Oct 08 00:15:00 crc kubenswrapper[4967]: I1008 00:15:00.398735 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f4296\" (UniqueName: \"kubernetes.io/projected/1f02cac1-b6b3-4a30-b2de-a2eb10a1f359-kube-api-access-f4296\") pod \"collect-profiles-29331375-vgl2p\" (UID: \"1f02cac1-b6b3-4a30-b2de-a2eb10a1f359\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331375-vgl2p" Oct 08 00:15:00 crc kubenswrapper[4967]: I1008 00:15:00.476158 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331375-vgl2p" Oct 08 00:15:00 crc kubenswrapper[4967]: I1008 00:15:00.706363 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331375-vgl2p"] Oct 08 00:15:00 crc kubenswrapper[4967]: W1008 00:15:00.717846 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1f02cac1_b6b3_4a30_b2de_a2eb10a1f359.slice/crio-965c3ce6550ea6436d24582f4aefc3bf604df401da407ddcc115ff9c4a2cca82 WatchSource:0}: Error finding container 965c3ce6550ea6436d24582f4aefc3bf604df401da407ddcc115ff9c4a2cca82: Status 404 returned error can't find the container with id 965c3ce6550ea6436d24582f4aefc3bf604df401da407ddcc115ff9c4a2cca82 Oct 08 00:15:01 crc kubenswrapper[4967]: I1008 00:15:01.052275 4967 generic.go:334] "Generic (PLEG): container finished" podID="1f02cac1-b6b3-4a30-b2de-a2eb10a1f359" containerID="3099e1417418bbdd481250f8ce2e3a6a050ab8e695a7012cfec629e0ed9af7cf" exitCode=0 Oct 08 00:15:01 crc kubenswrapper[4967]: I1008 00:15:01.052360 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331375-vgl2p" event={"ID":"1f02cac1-b6b3-4a30-b2de-a2eb10a1f359","Type":"ContainerDied","Data":"3099e1417418bbdd481250f8ce2e3a6a050ab8e695a7012cfec629e0ed9af7cf"} Oct 08 00:15:01 crc kubenswrapper[4967]: I1008 00:15:01.052574 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331375-vgl2p" event={"ID":"1f02cac1-b6b3-4a30-b2de-a2eb10a1f359","Type":"ContainerStarted","Data":"965c3ce6550ea6436d24582f4aefc3bf604df401da407ddcc115ff9c4a2cca82"} Oct 08 00:15:02 crc kubenswrapper[4967]: I1008 00:15:02.290538 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331375-vgl2p" Oct 08 00:15:02 crc kubenswrapper[4967]: I1008 00:15:02.394134 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1f02cac1-b6b3-4a30-b2de-a2eb10a1f359-config-volume\") pod \"1f02cac1-b6b3-4a30-b2de-a2eb10a1f359\" (UID: \"1f02cac1-b6b3-4a30-b2de-a2eb10a1f359\") " Oct 08 00:15:02 crc kubenswrapper[4967]: I1008 00:15:02.394176 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f4296\" (UniqueName: \"kubernetes.io/projected/1f02cac1-b6b3-4a30-b2de-a2eb10a1f359-kube-api-access-f4296\") pod \"1f02cac1-b6b3-4a30-b2de-a2eb10a1f359\" (UID: \"1f02cac1-b6b3-4a30-b2de-a2eb10a1f359\") " Oct 08 00:15:02 crc kubenswrapper[4967]: I1008 00:15:02.394239 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1f02cac1-b6b3-4a30-b2de-a2eb10a1f359-secret-volume\") pod \"1f02cac1-b6b3-4a30-b2de-a2eb10a1f359\" (UID: \"1f02cac1-b6b3-4a30-b2de-a2eb10a1f359\") " Oct 08 00:15:02 crc kubenswrapper[4967]: I1008 00:15:02.395606 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1f02cac1-b6b3-4a30-b2de-a2eb10a1f359-config-volume" (OuterVolumeSpecName: "config-volume") pod "1f02cac1-b6b3-4a30-b2de-a2eb10a1f359" (UID: "1f02cac1-b6b3-4a30-b2de-a2eb10a1f359"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:15:02 crc kubenswrapper[4967]: I1008 00:15:02.402011 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f02cac1-b6b3-4a30-b2de-a2eb10a1f359-kube-api-access-f4296" (OuterVolumeSpecName: "kube-api-access-f4296") pod "1f02cac1-b6b3-4a30-b2de-a2eb10a1f359" (UID: "1f02cac1-b6b3-4a30-b2de-a2eb10a1f359"). InnerVolumeSpecName "kube-api-access-f4296". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:15:02 crc kubenswrapper[4967]: I1008 00:15:02.402292 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f02cac1-b6b3-4a30-b2de-a2eb10a1f359-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "1f02cac1-b6b3-4a30-b2de-a2eb10a1f359" (UID: "1f02cac1-b6b3-4a30-b2de-a2eb10a1f359"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:15:02 crc kubenswrapper[4967]: I1008 00:15:02.495407 4967 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1f02cac1-b6b3-4a30-b2de-a2eb10a1f359-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 08 00:15:02 crc kubenswrapper[4967]: I1008 00:15:02.495463 4967 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1f02cac1-b6b3-4a30-b2de-a2eb10a1f359-config-volume\") on node \"crc\" DevicePath \"\"" Oct 08 00:15:02 crc kubenswrapper[4967]: I1008 00:15:02.495483 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f4296\" (UniqueName: \"kubernetes.io/projected/1f02cac1-b6b3-4a30-b2de-a2eb10a1f359-kube-api-access-f4296\") on node \"crc\" DevicePath \"\"" Oct 08 00:15:03 crc kubenswrapper[4967]: I1008 00:15:03.070106 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331375-vgl2p" event={"ID":"1f02cac1-b6b3-4a30-b2de-a2eb10a1f359","Type":"ContainerDied","Data":"965c3ce6550ea6436d24582f4aefc3bf604df401da407ddcc115ff9c4a2cca82"} Oct 08 00:15:03 crc kubenswrapper[4967]: I1008 00:15:03.070163 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="965c3ce6550ea6436d24582f4aefc3bf604df401da407ddcc115ff9c4a2cca82" Oct 08 00:15:03 crc kubenswrapper[4967]: I1008 00:15:03.070244 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331375-vgl2p" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.337976 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-bs8w9"] Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.338869 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" podUID="4f1af2ba-8907-4d14-afa1-ffde207c80f4" containerName="ovn-controller" containerID="cri-o://115f7a33f3dedae8f59d2f45f465f95f48b99c824e895ef7a0be9fe7a1b61d23" gracePeriod=30 Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.339244 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" podUID="4f1af2ba-8907-4d14-afa1-ffde207c80f4" containerName="sbdb" containerID="cri-o://27fb2b1225e386dcd7fecaaa0aa54ba1fdfafabfde6c3824ae07315313067542" gracePeriod=30 Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.339293 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" podUID="4f1af2ba-8907-4d14-afa1-ffde207c80f4" containerName="nbdb" containerID="cri-o://9c9fa0559eb35ba67e11e36069ac8d0b2fec76399788eef668a08ad4353df3ab" gracePeriod=30 Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.339326 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" podUID="4f1af2ba-8907-4d14-afa1-ffde207c80f4" containerName="northd" containerID="cri-o://d3312263e12bd42fbed0476ab5ace66b5497f62c9477865e130d67e3922e05cb" gracePeriod=30 Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.339356 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" podUID="4f1af2ba-8907-4d14-afa1-ffde207c80f4" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://b03b83ddff429dbfa8fe052901aaa86a39ff0dd75477ffa8a6b6f4198f5819ca" gracePeriod=30 Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.339383 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" podUID="4f1af2ba-8907-4d14-afa1-ffde207c80f4" containerName="kube-rbac-proxy-node" containerID="cri-o://4535cffb8c9d64e662e282ecb44f6817b4604866f0bb75015f3b888d5554f821" gracePeriod=30 Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.339438 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" podUID="4f1af2ba-8907-4d14-afa1-ffde207c80f4" containerName="ovn-acl-logging" containerID="cri-o://6e69dc603055aa221d44d11c049ddcc50cec568979911a310ed773a90341a499" gracePeriod=30 Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.381261 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" podUID="4f1af2ba-8907-4d14-afa1-ffde207c80f4" containerName="ovnkube-controller" containerID="cri-o://fab4c6588d3f2b44975eb2c248ad8b1cfef9b4b0fb797d0436123b63f9479df6" gracePeriod=30 Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.704038 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bs8w9_4f1af2ba-8907-4d14-afa1-ffde207c80f4/ovnkube-controller/3.log" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.708132 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bs8w9_4f1af2ba-8907-4d14-afa1-ffde207c80f4/ovn-acl-logging/0.log" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.709226 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bs8w9_4f1af2ba-8907-4d14-afa1-ffde207c80f4/ovn-controller/0.log" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.710120 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.776123 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-wdnnh"] Oct 08 00:15:40 crc kubenswrapper[4967]: E1008 00:15:40.776474 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f1af2ba-8907-4d14-afa1-ffde207c80f4" containerName="ovn-acl-logging" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.776498 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f1af2ba-8907-4d14-afa1-ffde207c80f4" containerName="ovn-acl-logging" Oct 08 00:15:40 crc kubenswrapper[4967]: E1008 00:15:40.776522 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f02cac1-b6b3-4a30-b2de-a2eb10a1f359" containerName="collect-profiles" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.776536 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f02cac1-b6b3-4a30-b2de-a2eb10a1f359" containerName="collect-profiles" Oct 08 00:15:40 crc kubenswrapper[4967]: E1008 00:15:40.776560 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f1af2ba-8907-4d14-afa1-ffde207c80f4" containerName="kubecfg-setup" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.776577 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f1af2ba-8907-4d14-afa1-ffde207c80f4" containerName="kubecfg-setup" Oct 08 00:15:40 crc kubenswrapper[4967]: E1008 00:15:40.776600 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f1af2ba-8907-4d14-afa1-ffde207c80f4" containerName="ovn-controller" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.776616 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f1af2ba-8907-4d14-afa1-ffde207c80f4" containerName="ovn-controller" Oct 08 00:15:40 crc kubenswrapper[4967]: E1008 00:15:40.776640 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f1af2ba-8907-4d14-afa1-ffde207c80f4" containerName="ovnkube-controller" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.776657 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f1af2ba-8907-4d14-afa1-ffde207c80f4" containerName="ovnkube-controller" Oct 08 00:15:40 crc kubenswrapper[4967]: E1008 00:15:40.776710 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f1af2ba-8907-4d14-afa1-ffde207c80f4" containerName="ovnkube-controller" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.776730 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f1af2ba-8907-4d14-afa1-ffde207c80f4" containerName="ovnkube-controller" Oct 08 00:15:40 crc kubenswrapper[4967]: E1008 00:15:40.776752 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f1af2ba-8907-4d14-afa1-ffde207c80f4" containerName="ovnkube-controller" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.776770 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f1af2ba-8907-4d14-afa1-ffde207c80f4" containerName="ovnkube-controller" Oct 08 00:15:40 crc kubenswrapper[4967]: E1008 00:15:40.776798 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f1af2ba-8907-4d14-afa1-ffde207c80f4" containerName="kube-rbac-proxy-ovn-metrics" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.776817 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f1af2ba-8907-4d14-afa1-ffde207c80f4" containerName="kube-rbac-proxy-ovn-metrics" Oct 08 00:15:40 crc kubenswrapper[4967]: E1008 00:15:40.776838 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f1af2ba-8907-4d14-afa1-ffde207c80f4" containerName="northd" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.776853 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f1af2ba-8907-4d14-afa1-ffde207c80f4" containerName="northd" Oct 08 00:15:40 crc kubenswrapper[4967]: E1008 00:15:40.776875 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f1af2ba-8907-4d14-afa1-ffde207c80f4" containerName="kube-rbac-proxy-node" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.776895 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f1af2ba-8907-4d14-afa1-ffde207c80f4" containerName="kube-rbac-proxy-node" Oct 08 00:15:40 crc kubenswrapper[4967]: E1008 00:15:40.776917 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f1af2ba-8907-4d14-afa1-ffde207c80f4" containerName="nbdb" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.776935 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f1af2ba-8907-4d14-afa1-ffde207c80f4" containerName="nbdb" Oct 08 00:15:40 crc kubenswrapper[4967]: E1008 00:15:40.776957 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f1af2ba-8907-4d14-afa1-ffde207c80f4" containerName="sbdb" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.776970 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f1af2ba-8907-4d14-afa1-ffde207c80f4" containerName="sbdb" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.777225 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f1af2ba-8907-4d14-afa1-ffde207c80f4" containerName="ovnkube-controller" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.777253 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f1af2ba-8907-4d14-afa1-ffde207c80f4" containerName="ovnkube-controller" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.777270 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f1af2ba-8907-4d14-afa1-ffde207c80f4" containerName="ovn-acl-logging" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.777299 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f1af2ba-8907-4d14-afa1-ffde207c80f4" containerName="kube-rbac-proxy-node" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.777323 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f1af2ba-8907-4d14-afa1-ffde207c80f4" containerName="ovnkube-controller" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.777343 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f02cac1-b6b3-4a30-b2de-a2eb10a1f359" containerName="collect-profiles" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.777362 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f1af2ba-8907-4d14-afa1-ffde207c80f4" containerName="ovnkube-controller" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.777386 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f1af2ba-8907-4d14-afa1-ffde207c80f4" containerName="sbdb" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.777412 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f1af2ba-8907-4d14-afa1-ffde207c80f4" containerName="nbdb" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.777431 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f1af2ba-8907-4d14-afa1-ffde207c80f4" containerName="ovn-controller" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.777457 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f1af2ba-8907-4d14-afa1-ffde207c80f4" containerName="kube-rbac-proxy-ovn-metrics" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.777481 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f1af2ba-8907-4d14-afa1-ffde207c80f4" containerName="northd" Oct 08 00:15:40 crc kubenswrapper[4967]: E1008 00:15:40.777718 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f1af2ba-8907-4d14-afa1-ffde207c80f4" containerName="ovnkube-controller" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.777756 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f1af2ba-8907-4d14-afa1-ffde207c80f4" containerName="ovnkube-controller" Oct 08 00:15:40 crc kubenswrapper[4967]: E1008 00:15:40.777788 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f1af2ba-8907-4d14-afa1-ffde207c80f4" containerName="ovnkube-controller" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.777806 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f1af2ba-8907-4d14-afa1-ffde207c80f4" containerName="ovnkube-controller" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.778053 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f1af2ba-8907-4d14-afa1-ffde207c80f4" containerName="ovnkube-controller" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.781796 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-wdnnh" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.820259 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-log-socket\") pod \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\" (UID: \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\") " Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.820314 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-host-var-lib-cni-networks-ovn-kubernetes\") pod \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\" (UID: \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\") " Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.820360 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/4f1af2ba-8907-4d14-afa1-ffde207c80f4-ovnkube-script-lib\") pod \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\" (UID: \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\") " Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.820401 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-run-ovn\") pod \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\" (UID: \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\") " Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.820422 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-log-socket" (OuterVolumeSpecName: "log-socket") pod "4f1af2ba-8907-4d14-afa1-ffde207c80f4" (UID: "4f1af2ba-8907-4d14-afa1-ffde207c80f4"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.820449 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-host-run-netns\") pod \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\" (UID: \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\") " Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.820477 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "4f1af2ba-8907-4d14-afa1-ffde207c80f4" (UID: "4f1af2ba-8907-4d14-afa1-ffde207c80f4"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.820481 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-host-cni-bin\") pod \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\" (UID: \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\") " Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.820486 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "4f1af2ba-8907-4d14-afa1-ffde207c80f4" (UID: "4f1af2ba-8907-4d14-afa1-ffde207c80f4"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.820519 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "4f1af2ba-8907-4d14-afa1-ffde207c80f4" (UID: "4f1af2ba-8907-4d14-afa1-ffde207c80f4"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.820525 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-etc-openvswitch\") pod \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\" (UID: \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\") " Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.820559 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "4f1af2ba-8907-4d14-afa1-ffde207c80f4" (UID: "4f1af2ba-8907-4d14-afa1-ffde207c80f4"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.820570 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/4f1af2ba-8907-4d14-afa1-ffde207c80f4-env-overrides\") pod \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\" (UID: \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\") " Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.820597 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-run-openvswitch\") pod \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\" (UID: \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\") " Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.820608 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "4f1af2ba-8907-4d14-afa1-ffde207c80f4" (UID: "4f1af2ba-8907-4d14-afa1-ffde207c80f4"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.820654 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "4f1af2ba-8907-4d14-afa1-ffde207c80f4" (UID: "4f1af2ba-8907-4d14-afa1-ffde207c80f4"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.820626 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-host-cni-netd\") pod \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\" (UID: \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\") " Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.820719 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "4f1af2ba-8907-4d14-afa1-ffde207c80f4" (UID: "4f1af2ba-8907-4d14-afa1-ffde207c80f4"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.820736 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-host-kubelet\") pod \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\" (UID: \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\") " Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.820784 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-var-lib-openvswitch\") pod \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\" (UID: \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\") " Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.820821 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qlrlf\" (UniqueName: \"kubernetes.io/projected/4f1af2ba-8907-4d14-afa1-ffde207c80f4-kube-api-access-qlrlf\") pod \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\" (UID: \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\") " Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.820837 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "4f1af2ba-8907-4d14-afa1-ffde207c80f4" (UID: "4f1af2ba-8907-4d14-afa1-ffde207c80f4"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.820864 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "4f1af2ba-8907-4d14-afa1-ffde207c80f4" (UID: "4f1af2ba-8907-4d14-afa1-ffde207c80f4"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.820871 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-host-run-ovn-kubernetes\") pod \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\" (UID: \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\") " Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.820903 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "4f1af2ba-8907-4d14-afa1-ffde207c80f4" (UID: "4f1af2ba-8907-4d14-afa1-ffde207c80f4"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.820964 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-host-slash\") pod \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\" (UID: \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\") " Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.820999 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-run-systemd\") pod \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\" (UID: \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\") " Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.821025 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-node-log\") pod \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\" (UID: \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\") " Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.821052 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-systemd-units\") pod \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\" (UID: \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\") " Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.821059 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-host-slash" (OuterVolumeSpecName: "host-slash") pod "4f1af2ba-8907-4d14-afa1-ffde207c80f4" (UID: "4f1af2ba-8907-4d14-afa1-ffde207c80f4"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.821077 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4f1af2ba-8907-4d14-afa1-ffde207c80f4-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "4f1af2ba-8907-4d14-afa1-ffde207c80f4" (UID: "4f1af2ba-8907-4d14-afa1-ffde207c80f4"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.821092 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/4f1af2ba-8907-4d14-afa1-ffde207c80f4-ovn-node-metrics-cert\") pod \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\" (UID: \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\") " Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.821104 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-node-log" (OuterVolumeSpecName: "node-log") pod "4f1af2ba-8907-4d14-afa1-ffde207c80f4" (UID: "4f1af2ba-8907-4d14-afa1-ffde207c80f4"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.821101 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4f1af2ba-8907-4d14-afa1-ffde207c80f4-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "4f1af2ba-8907-4d14-afa1-ffde207c80f4" (UID: "4f1af2ba-8907-4d14-afa1-ffde207c80f4"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.821124 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/4f1af2ba-8907-4d14-afa1-ffde207c80f4-ovnkube-config\") pod \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\" (UID: \"4f1af2ba-8907-4d14-afa1-ffde207c80f4\") " Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.821128 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "4f1af2ba-8907-4d14-afa1-ffde207c80f4" (UID: "4f1af2ba-8907-4d14-afa1-ffde207c80f4"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.821415 4967 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.821439 4967 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-host-slash\") on node \"crc\" DevicePath \"\"" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.821455 4967 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-node-log\") on node \"crc\" DevicePath \"\"" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.821469 4967 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-systemd-units\") on node \"crc\" DevicePath \"\"" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.821483 4967 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-log-socket\") on node \"crc\" DevicePath \"\"" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.821498 4967 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/4f1af2ba-8907-4d14-afa1-ffde207c80f4-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.821516 4967 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.821533 4967 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.821535 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4f1af2ba-8907-4d14-afa1-ffde207c80f4-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "4f1af2ba-8907-4d14-afa1-ffde207c80f4" (UID: "4f1af2ba-8907-4d14-afa1-ffde207c80f4"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.821548 4967 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-host-run-netns\") on node \"crc\" DevicePath \"\"" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.821563 4967 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.821578 4967 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-host-cni-bin\") on node \"crc\" DevicePath \"\"" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.821593 4967 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-run-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.821605 4967 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-host-cni-netd\") on node \"crc\" DevicePath \"\"" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.821615 4967 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/4f1af2ba-8907-4d14-afa1-ffde207c80f4-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.821626 4967 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-host-kubelet\") on node \"crc\" DevicePath \"\"" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.821636 4967 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.827028 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f1af2ba-8907-4d14-afa1-ffde207c80f4-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "4f1af2ba-8907-4d14-afa1-ffde207c80f4" (UID: "4f1af2ba-8907-4d14-afa1-ffde207c80f4"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.827996 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f1af2ba-8907-4d14-afa1-ffde207c80f4-kube-api-access-qlrlf" (OuterVolumeSpecName: "kube-api-access-qlrlf") pod "4f1af2ba-8907-4d14-afa1-ffde207c80f4" (UID: "4f1af2ba-8907-4d14-afa1-ffde207c80f4"). InnerVolumeSpecName "kube-api-access-qlrlf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.837073 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "4f1af2ba-8907-4d14-afa1-ffde207c80f4" (UID: "4f1af2ba-8907-4d14-afa1-ffde207c80f4"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.923107 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4da518e4-d3e3-437c-baa8-4fdef073dde5-var-lib-openvswitch\") pod \"ovnkube-node-wdnnh\" (UID: \"4da518e4-d3e3-437c-baa8-4fdef073dde5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdnnh" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.923218 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/4da518e4-d3e3-437c-baa8-4fdef073dde5-node-log\") pod \"ovnkube-node-wdnnh\" (UID: \"4da518e4-d3e3-437c-baa8-4fdef073dde5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdnnh" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.923245 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/4da518e4-d3e3-437c-baa8-4fdef073dde5-ovnkube-script-lib\") pod \"ovnkube-node-wdnnh\" (UID: \"4da518e4-d3e3-437c-baa8-4fdef073dde5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdnnh" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.923387 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/4da518e4-d3e3-437c-baa8-4fdef073dde5-host-kubelet\") pod \"ovnkube-node-wdnnh\" (UID: \"4da518e4-d3e3-437c-baa8-4fdef073dde5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdnnh" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.923473 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/4da518e4-d3e3-437c-baa8-4fdef073dde5-ovnkube-config\") pod \"ovnkube-node-wdnnh\" (UID: \"4da518e4-d3e3-437c-baa8-4fdef073dde5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdnnh" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.923517 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/4da518e4-d3e3-437c-baa8-4fdef073dde5-host-slash\") pod \"ovnkube-node-wdnnh\" (UID: \"4da518e4-d3e3-437c-baa8-4fdef073dde5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdnnh" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.923560 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/4da518e4-d3e3-437c-baa8-4fdef073dde5-run-systemd\") pod \"ovnkube-node-wdnnh\" (UID: \"4da518e4-d3e3-437c-baa8-4fdef073dde5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdnnh" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.923600 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4da518e4-d3e3-437c-baa8-4fdef073dde5-host-run-netns\") pod \"ovnkube-node-wdnnh\" (UID: \"4da518e4-d3e3-437c-baa8-4fdef073dde5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdnnh" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.923659 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4da518e4-d3e3-437c-baa8-4fdef073dde5-etc-openvswitch\") pod \"ovnkube-node-wdnnh\" (UID: \"4da518e4-d3e3-437c-baa8-4fdef073dde5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdnnh" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.923770 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4da518e4-d3e3-437c-baa8-4fdef073dde5-host-run-ovn-kubernetes\") pod \"ovnkube-node-wdnnh\" (UID: \"4da518e4-d3e3-437c-baa8-4fdef073dde5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdnnh" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.923819 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4da518e4-d3e3-437c-baa8-4fdef073dde5-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-wdnnh\" (UID: \"4da518e4-d3e3-437c-baa8-4fdef073dde5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdnnh" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.923992 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/4da518e4-d3e3-437c-baa8-4fdef073dde5-log-socket\") pod \"ovnkube-node-wdnnh\" (UID: \"4da518e4-d3e3-437c-baa8-4fdef073dde5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdnnh" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.924034 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/4da518e4-d3e3-437c-baa8-4fdef073dde5-ovn-node-metrics-cert\") pod \"ovnkube-node-wdnnh\" (UID: \"4da518e4-d3e3-437c-baa8-4fdef073dde5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdnnh" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.924065 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4da518e4-d3e3-437c-baa8-4fdef073dde5-run-openvswitch\") pod \"ovnkube-node-wdnnh\" (UID: \"4da518e4-d3e3-437c-baa8-4fdef073dde5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdnnh" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.924093 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/4da518e4-d3e3-437c-baa8-4fdef073dde5-systemd-units\") pod \"ovnkube-node-wdnnh\" (UID: \"4da518e4-d3e3-437c-baa8-4fdef073dde5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdnnh" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.924120 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/4da518e4-d3e3-437c-baa8-4fdef073dde5-env-overrides\") pod \"ovnkube-node-wdnnh\" (UID: \"4da518e4-d3e3-437c-baa8-4fdef073dde5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdnnh" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.924146 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/4da518e4-d3e3-437c-baa8-4fdef073dde5-run-ovn\") pod \"ovnkube-node-wdnnh\" (UID: \"4da518e4-d3e3-437c-baa8-4fdef073dde5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdnnh" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.924191 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4da518e4-d3e3-437c-baa8-4fdef073dde5-host-cni-bin\") pod \"ovnkube-node-wdnnh\" (UID: \"4da518e4-d3e3-437c-baa8-4fdef073dde5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdnnh" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.924216 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/4da518e4-d3e3-437c-baa8-4fdef073dde5-host-cni-netd\") pod \"ovnkube-node-wdnnh\" (UID: \"4da518e4-d3e3-437c-baa8-4fdef073dde5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdnnh" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.924236 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2vftr\" (UniqueName: \"kubernetes.io/projected/4da518e4-d3e3-437c-baa8-4fdef073dde5-kube-api-access-2vftr\") pod \"ovnkube-node-wdnnh\" (UID: \"4da518e4-d3e3-437c-baa8-4fdef073dde5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdnnh" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.924295 4967 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/4f1af2ba-8907-4d14-afa1-ffde207c80f4-run-systemd\") on node \"crc\" DevicePath \"\"" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.924317 4967 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/4f1af2ba-8907-4d14-afa1-ffde207c80f4-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.924332 4967 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/4f1af2ba-8907-4d14-afa1-ffde207c80f4-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 08 00:15:40 crc kubenswrapper[4967]: I1008 00:15:40.924345 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qlrlf\" (UniqueName: \"kubernetes.io/projected/4f1af2ba-8907-4d14-afa1-ffde207c80f4-kube-api-access-qlrlf\") on node \"crc\" DevicePath \"\"" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.026314 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4da518e4-d3e3-437c-baa8-4fdef073dde5-host-cni-bin\") pod \"ovnkube-node-wdnnh\" (UID: \"4da518e4-d3e3-437c-baa8-4fdef073dde5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdnnh" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.026392 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/4da518e4-d3e3-437c-baa8-4fdef073dde5-host-cni-netd\") pod \"ovnkube-node-wdnnh\" (UID: \"4da518e4-d3e3-437c-baa8-4fdef073dde5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdnnh" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.026494 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2vftr\" (UniqueName: \"kubernetes.io/projected/4da518e4-d3e3-437c-baa8-4fdef073dde5-kube-api-access-2vftr\") pod \"ovnkube-node-wdnnh\" (UID: \"4da518e4-d3e3-437c-baa8-4fdef073dde5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdnnh" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.026559 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/4da518e4-d3e3-437c-baa8-4fdef073dde5-host-cni-netd\") pod \"ovnkube-node-wdnnh\" (UID: \"4da518e4-d3e3-437c-baa8-4fdef073dde5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdnnh" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.026582 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4da518e4-d3e3-437c-baa8-4fdef073dde5-var-lib-openvswitch\") pod \"ovnkube-node-wdnnh\" (UID: \"4da518e4-d3e3-437c-baa8-4fdef073dde5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdnnh" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.026484 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4da518e4-d3e3-437c-baa8-4fdef073dde5-host-cni-bin\") pod \"ovnkube-node-wdnnh\" (UID: \"4da518e4-d3e3-437c-baa8-4fdef073dde5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdnnh" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.026716 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/4da518e4-d3e3-437c-baa8-4fdef073dde5-node-log\") pod \"ovnkube-node-wdnnh\" (UID: \"4da518e4-d3e3-437c-baa8-4fdef073dde5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdnnh" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.026732 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4da518e4-d3e3-437c-baa8-4fdef073dde5-var-lib-openvswitch\") pod \"ovnkube-node-wdnnh\" (UID: \"4da518e4-d3e3-437c-baa8-4fdef073dde5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdnnh" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.026782 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/4da518e4-d3e3-437c-baa8-4fdef073dde5-ovnkube-script-lib\") pod \"ovnkube-node-wdnnh\" (UID: \"4da518e4-d3e3-437c-baa8-4fdef073dde5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdnnh" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.026826 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/4da518e4-d3e3-437c-baa8-4fdef073dde5-node-log\") pod \"ovnkube-node-wdnnh\" (UID: \"4da518e4-d3e3-437c-baa8-4fdef073dde5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdnnh" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.026855 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/4da518e4-d3e3-437c-baa8-4fdef073dde5-host-kubelet\") pod \"ovnkube-node-wdnnh\" (UID: \"4da518e4-d3e3-437c-baa8-4fdef073dde5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdnnh" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.026906 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/4da518e4-d3e3-437c-baa8-4fdef073dde5-host-kubelet\") pod \"ovnkube-node-wdnnh\" (UID: \"4da518e4-d3e3-437c-baa8-4fdef073dde5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdnnh" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.026931 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/4da518e4-d3e3-437c-baa8-4fdef073dde5-ovnkube-config\") pod \"ovnkube-node-wdnnh\" (UID: \"4da518e4-d3e3-437c-baa8-4fdef073dde5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdnnh" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.026985 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/4da518e4-d3e3-437c-baa8-4fdef073dde5-host-slash\") pod \"ovnkube-node-wdnnh\" (UID: \"4da518e4-d3e3-437c-baa8-4fdef073dde5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdnnh" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.027056 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/4da518e4-d3e3-437c-baa8-4fdef073dde5-host-slash\") pod \"ovnkube-node-wdnnh\" (UID: \"4da518e4-d3e3-437c-baa8-4fdef073dde5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdnnh" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.027065 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/4da518e4-d3e3-437c-baa8-4fdef073dde5-run-systemd\") pod \"ovnkube-node-wdnnh\" (UID: \"4da518e4-d3e3-437c-baa8-4fdef073dde5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdnnh" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.027116 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4da518e4-d3e3-437c-baa8-4fdef073dde5-host-run-netns\") pod \"ovnkube-node-wdnnh\" (UID: \"4da518e4-d3e3-437c-baa8-4fdef073dde5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdnnh" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.027166 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4da518e4-d3e3-437c-baa8-4fdef073dde5-etc-openvswitch\") pod \"ovnkube-node-wdnnh\" (UID: \"4da518e4-d3e3-437c-baa8-4fdef073dde5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdnnh" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.027180 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4da518e4-d3e3-437c-baa8-4fdef073dde5-host-run-netns\") pod \"ovnkube-node-wdnnh\" (UID: \"4da518e4-d3e3-437c-baa8-4fdef073dde5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdnnh" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.027195 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/4da518e4-d3e3-437c-baa8-4fdef073dde5-run-systemd\") pod \"ovnkube-node-wdnnh\" (UID: \"4da518e4-d3e3-437c-baa8-4fdef073dde5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdnnh" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.027240 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4da518e4-d3e3-437c-baa8-4fdef073dde5-etc-openvswitch\") pod \"ovnkube-node-wdnnh\" (UID: \"4da518e4-d3e3-437c-baa8-4fdef073dde5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdnnh" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.027244 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4da518e4-d3e3-437c-baa8-4fdef073dde5-host-run-ovn-kubernetes\") pod \"ovnkube-node-wdnnh\" (UID: \"4da518e4-d3e3-437c-baa8-4fdef073dde5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdnnh" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.027296 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4da518e4-d3e3-437c-baa8-4fdef073dde5-host-run-ovn-kubernetes\") pod \"ovnkube-node-wdnnh\" (UID: \"4da518e4-d3e3-437c-baa8-4fdef073dde5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdnnh" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.027349 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4da518e4-d3e3-437c-baa8-4fdef073dde5-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-wdnnh\" (UID: \"4da518e4-d3e3-437c-baa8-4fdef073dde5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdnnh" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.027313 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4da518e4-d3e3-437c-baa8-4fdef073dde5-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-wdnnh\" (UID: \"4da518e4-d3e3-437c-baa8-4fdef073dde5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdnnh" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.027454 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/4da518e4-d3e3-437c-baa8-4fdef073dde5-log-socket\") pod \"ovnkube-node-wdnnh\" (UID: \"4da518e4-d3e3-437c-baa8-4fdef073dde5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdnnh" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.027509 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/4da518e4-d3e3-437c-baa8-4fdef073dde5-ovn-node-metrics-cert\") pod \"ovnkube-node-wdnnh\" (UID: \"4da518e4-d3e3-437c-baa8-4fdef073dde5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdnnh" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.027556 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/4da518e4-d3e3-437c-baa8-4fdef073dde5-log-socket\") pod \"ovnkube-node-wdnnh\" (UID: \"4da518e4-d3e3-437c-baa8-4fdef073dde5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdnnh" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.027568 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4da518e4-d3e3-437c-baa8-4fdef073dde5-run-openvswitch\") pod \"ovnkube-node-wdnnh\" (UID: \"4da518e4-d3e3-437c-baa8-4fdef073dde5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdnnh" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.027618 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/4da518e4-d3e3-437c-baa8-4fdef073dde5-systemd-units\") pod \"ovnkube-node-wdnnh\" (UID: \"4da518e4-d3e3-437c-baa8-4fdef073dde5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdnnh" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.027654 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/4da518e4-d3e3-437c-baa8-4fdef073dde5-run-openvswitch\") pod \"ovnkube-node-wdnnh\" (UID: \"4da518e4-d3e3-437c-baa8-4fdef073dde5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdnnh" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.027713 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/4da518e4-d3e3-437c-baa8-4fdef073dde5-run-ovn\") pod \"ovnkube-node-wdnnh\" (UID: \"4da518e4-d3e3-437c-baa8-4fdef073dde5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdnnh" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.027767 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/4da518e4-d3e3-437c-baa8-4fdef073dde5-run-ovn\") pod \"ovnkube-node-wdnnh\" (UID: \"4da518e4-d3e3-437c-baa8-4fdef073dde5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdnnh" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.027771 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/4da518e4-d3e3-437c-baa8-4fdef073dde5-env-overrides\") pod \"ovnkube-node-wdnnh\" (UID: \"4da518e4-d3e3-437c-baa8-4fdef073dde5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdnnh" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.027725 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/4da518e4-d3e3-437c-baa8-4fdef073dde5-systemd-units\") pod \"ovnkube-node-wdnnh\" (UID: \"4da518e4-d3e3-437c-baa8-4fdef073dde5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdnnh" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.028200 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/4da518e4-d3e3-437c-baa8-4fdef073dde5-ovnkube-script-lib\") pod \"ovnkube-node-wdnnh\" (UID: \"4da518e4-d3e3-437c-baa8-4fdef073dde5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdnnh" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.028522 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/4da518e4-d3e3-437c-baa8-4fdef073dde5-ovnkube-config\") pod \"ovnkube-node-wdnnh\" (UID: \"4da518e4-d3e3-437c-baa8-4fdef073dde5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdnnh" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.028715 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/4da518e4-d3e3-437c-baa8-4fdef073dde5-env-overrides\") pod \"ovnkube-node-wdnnh\" (UID: \"4da518e4-d3e3-437c-baa8-4fdef073dde5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdnnh" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.032832 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/4da518e4-d3e3-437c-baa8-4fdef073dde5-ovn-node-metrics-cert\") pod \"ovnkube-node-wdnnh\" (UID: \"4da518e4-d3e3-437c-baa8-4fdef073dde5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdnnh" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.051088 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2vftr\" (UniqueName: \"kubernetes.io/projected/4da518e4-d3e3-437c-baa8-4fdef073dde5-kube-api-access-2vftr\") pod \"ovnkube-node-wdnnh\" (UID: \"4da518e4-d3e3-437c-baa8-4fdef073dde5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wdnnh" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.105761 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-wdnnh" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.332382 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bs8w9_4f1af2ba-8907-4d14-afa1-ffde207c80f4/ovnkube-controller/3.log" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.339092 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bs8w9_4f1af2ba-8907-4d14-afa1-ffde207c80f4/ovn-acl-logging/0.log" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.348517 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bs8w9_4f1af2ba-8907-4d14-afa1-ffde207c80f4/ovn-controller/0.log" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.349804 4967 generic.go:334] "Generic (PLEG): container finished" podID="4f1af2ba-8907-4d14-afa1-ffde207c80f4" containerID="fab4c6588d3f2b44975eb2c248ad8b1cfef9b4b0fb797d0436123b63f9479df6" exitCode=0 Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.349862 4967 generic.go:334] "Generic (PLEG): container finished" podID="4f1af2ba-8907-4d14-afa1-ffde207c80f4" containerID="27fb2b1225e386dcd7fecaaa0aa54ba1fdfafabfde6c3824ae07315313067542" exitCode=0 Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.349888 4967 generic.go:334] "Generic (PLEG): container finished" podID="4f1af2ba-8907-4d14-afa1-ffde207c80f4" containerID="9c9fa0559eb35ba67e11e36069ac8d0b2fec76399788eef668a08ad4353df3ab" exitCode=0 Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.349907 4967 generic.go:334] "Generic (PLEG): container finished" podID="4f1af2ba-8907-4d14-afa1-ffde207c80f4" containerID="d3312263e12bd42fbed0476ab5ace66b5497f62c9477865e130d67e3922e05cb" exitCode=0 Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.349925 4967 generic.go:334] "Generic (PLEG): container finished" podID="4f1af2ba-8907-4d14-afa1-ffde207c80f4" containerID="b03b83ddff429dbfa8fe052901aaa86a39ff0dd75477ffa8a6b6f4198f5819ca" exitCode=0 Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.349936 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.349979 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" event={"ID":"4f1af2ba-8907-4d14-afa1-ffde207c80f4","Type":"ContainerDied","Data":"fab4c6588d3f2b44975eb2c248ad8b1cfef9b4b0fb797d0436123b63f9479df6"} Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.350067 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" event={"ID":"4f1af2ba-8907-4d14-afa1-ffde207c80f4","Type":"ContainerDied","Data":"27fb2b1225e386dcd7fecaaa0aa54ba1fdfafabfde6c3824ae07315313067542"} Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.349943 4967 generic.go:334] "Generic (PLEG): container finished" podID="4f1af2ba-8907-4d14-afa1-ffde207c80f4" containerID="4535cffb8c9d64e662e282ecb44f6817b4604866f0bb75015f3b888d5554f821" exitCode=0 Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.350125 4967 scope.go:117] "RemoveContainer" containerID="fab4c6588d3f2b44975eb2c248ad8b1cfef9b4b0fb797d0436123b63f9479df6" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.350184 4967 generic.go:334] "Generic (PLEG): container finished" podID="4f1af2ba-8907-4d14-afa1-ffde207c80f4" containerID="6e69dc603055aa221d44d11c049ddcc50cec568979911a310ed773a90341a499" exitCode=143 Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.350229 4967 generic.go:334] "Generic (PLEG): container finished" podID="4f1af2ba-8907-4d14-afa1-ffde207c80f4" containerID="115f7a33f3dedae8f59d2f45f465f95f48b99c824e895ef7a0be9fe7a1b61d23" exitCode=143 Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.350102 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" event={"ID":"4f1af2ba-8907-4d14-afa1-ffde207c80f4","Type":"ContainerDied","Data":"9c9fa0559eb35ba67e11e36069ac8d0b2fec76399788eef668a08ad4353df3ab"} Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.350374 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" event={"ID":"4f1af2ba-8907-4d14-afa1-ffde207c80f4","Type":"ContainerDied","Data":"d3312263e12bd42fbed0476ab5ace66b5497f62c9477865e130d67e3922e05cb"} Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.350421 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" event={"ID":"4f1af2ba-8907-4d14-afa1-ffde207c80f4","Type":"ContainerDied","Data":"b03b83ddff429dbfa8fe052901aaa86a39ff0dd75477ffa8a6b6f4198f5819ca"} Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.350451 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" event={"ID":"4f1af2ba-8907-4d14-afa1-ffde207c80f4","Type":"ContainerDied","Data":"4535cffb8c9d64e662e282ecb44f6817b4604866f0bb75015f3b888d5554f821"} Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.350478 4967 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4f7d46ee97fd3c4286f45879fededea788052cf83ebeebcb539be452f2ae102d"} Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.350503 4967 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"27fb2b1225e386dcd7fecaaa0aa54ba1fdfafabfde6c3824ae07315313067542"} Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.350520 4967 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9c9fa0559eb35ba67e11e36069ac8d0b2fec76399788eef668a08ad4353df3ab"} Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.350535 4967 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d3312263e12bd42fbed0476ab5ace66b5497f62c9477865e130d67e3922e05cb"} Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.350551 4967 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b03b83ddff429dbfa8fe052901aaa86a39ff0dd75477ffa8a6b6f4198f5819ca"} Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.350568 4967 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4535cffb8c9d64e662e282ecb44f6817b4604866f0bb75015f3b888d5554f821"} Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.350586 4967 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6e69dc603055aa221d44d11c049ddcc50cec568979911a310ed773a90341a499"} Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.350602 4967 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"115f7a33f3dedae8f59d2f45f465f95f48b99c824e895ef7a0be9fe7a1b61d23"} Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.350617 4967 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427"} Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.350638 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" event={"ID":"4f1af2ba-8907-4d14-afa1-ffde207c80f4","Type":"ContainerDied","Data":"6e69dc603055aa221d44d11c049ddcc50cec568979911a310ed773a90341a499"} Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.350663 4967 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fab4c6588d3f2b44975eb2c248ad8b1cfef9b4b0fb797d0436123b63f9479df6"} Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.350713 4967 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4f7d46ee97fd3c4286f45879fededea788052cf83ebeebcb539be452f2ae102d"} Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.350730 4967 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"27fb2b1225e386dcd7fecaaa0aa54ba1fdfafabfde6c3824ae07315313067542"} Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.350746 4967 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9c9fa0559eb35ba67e11e36069ac8d0b2fec76399788eef668a08ad4353df3ab"} Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.350761 4967 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d3312263e12bd42fbed0476ab5ace66b5497f62c9477865e130d67e3922e05cb"} Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.350777 4967 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b03b83ddff429dbfa8fe052901aaa86a39ff0dd75477ffa8a6b6f4198f5819ca"} Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.350792 4967 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4535cffb8c9d64e662e282ecb44f6817b4604866f0bb75015f3b888d5554f821"} Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.350807 4967 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6e69dc603055aa221d44d11c049ddcc50cec568979911a310ed773a90341a499"} Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.350847 4967 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"115f7a33f3dedae8f59d2f45f465f95f48b99c824e895ef7a0be9fe7a1b61d23"} Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.350863 4967 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427"} Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.350907 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" event={"ID":"4f1af2ba-8907-4d14-afa1-ffde207c80f4","Type":"ContainerDied","Data":"115f7a33f3dedae8f59d2f45f465f95f48b99c824e895ef7a0be9fe7a1b61d23"} Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.350932 4967 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fab4c6588d3f2b44975eb2c248ad8b1cfef9b4b0fb797d0436123b63f9479df6"} Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.350949 4967 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4f7d46ee97fd3c4286f45879fededea788052cf83ebeebcb539be452f2ae102d"} Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.350965 4967 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"27fb2b1225e386dcd7fecaaa0aa54ba1fdfafabfde6c3824ae07315313067542"} Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.350981 4967 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9c9fa0559eb35ba67e11e36069ac8d0b2fec76399788eef668a08ad4353df3ab"} Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.350997 4967 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d3312263e12bd42fbed0476ab5ace66b5497f62c9477865e130d67e3922e05cb"} Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.351012 4967 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b03b83ddff429dbfa8fe052901aaa86a39ff0dd75477ffa8a6b6f4198f5819ca"} Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.351027 4967 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4535cffb8c9d64e662e282ecb44f6817b4604866f0bb75015f3b888d5554f821"} Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.351042 4967 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6e69dc603055aa221d44d11c049ddcc50cec568979911a310ed773a90341a499"} Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.351057 4967 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"115f7a33f3dedae8f59d2f45f465f95f48b99c824e895ef7a0be9fe7a1b61d23"} Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.351074 4967 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427"} Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.351096 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bs8w9" event={"ID":"4f1af2ba-8907-4d14-afa1-ffde207c80f4","Type":"ContainerDied","Data":"11f2eddeaf743f73001e819048455fe3ce59f7dbad311bd1525a412bf99ff182"} Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.351122 4967 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fab4c6588d3f2b44975eb2c248ad8b1cfef9b4b0fb797d0436123b63f9479df6"} Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.351140 4967 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4f7d46ee97fd3c4286f45879fededea788052cf83ebeebcb539be452f2ae102d"} Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.351157 4967 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"27fb2b1225e386dcd7fecaaa0aa54ba1fdfafabfde6c3824ae07315313067542"} Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.351172 4967 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9c9fa0559eb35ba67e11e36069ac8d0b2fec76399788eef668a08ad4353df3ab"} Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.351187 4967 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d3312263e12bd42fbed0476ab5ace66b5497f62c9477865e130d67e3922e05cb"} Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.351203 4967 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b03b83ddff429dbfa8fe052901aaa86a39ff0dd75477ffa8a6b6f4198f5819ca"} Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.351217 4967 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"4535cffb8c9d64e662e282ecb44f6817b4604866f0bb75015f3b888d5554f821"} Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.351231 4967 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6e69dc603055aa221d44d11c049ddcc50cec568979911a310ed773a90341a499"} Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.351247 4967 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"115f7a33f3dedae8f59d2f45f465f95f48b99c824e895ef7a0be9fe7a1b61d23"} Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.351261 4967 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427"} Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.360786 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-npz2s_c71501b1-aa9e-420e-b1bd-7e15365e7adc/kube-multus/1.log" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.361990 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-npz2s_c71501b1-aa9e-420e-b1bd-7e15365e7adc/kube-multus/0.log" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.362047 4967 generic.go:334] "Generic (PLEG): container finished" podID="c71501b1-aa9e-420e-b1bd-7e15365e7adc" containerID="9010a5d6165a64f986f9101dbbbcaf8a7d3fa5af85c14be4bf3aa8ed8536ad4e" exitCode=2 Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.362124 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-npz2s" event={"ID":"c71501b1-aa9e-420e-b1bd-7e15365e7adc","Type":"ContainerDied","Data":"9010a5d6165a64f986f9101dbbbcaf8a7d3fa5af85c14be4bf3aa8ed8536ad4e"} Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.362161 4967 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"82dfe6c420d804b2d7daef4b5cfe3f87b77b369287c4d65553fa93f5f24f589d"} Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.362940 4967 scope.go:117] "RemoveContainer" containerID="9010a5d6165a64f986f9101dbbbcaf8a7d3fa5af85c14be4bf3aa8ed8536ad4e" Oct 08 00:15:41 crc kubenswrapper[4967]: E1008 00:15:41.363254 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-npz2s_openshift-multus(c71501b1-aa9e-420e-b1bd-7e15365e7adc)\"" pod="openshift-multus/multus-npz2s" podUID="c71501b1-aa9e-420e-b1bd-7e15365e7adc" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.367943 4967 generic.go:334] "Generic (PLEG): container finished" podID="4da518e4-d3e3-437c-baa8-4fdef073dde5" containerID="f06102f9d44d04b77d72a30ac1c06af34647db2c3c607cfceb7ee637622e3f22" exitCode=0 Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.368028 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wdnnh" event={"ID":"4da518e4-d3e3-437c-baa8-4fdef073dde5","Type":"ContainerDied","Data":"f06102f9d44d04b77d72a30ac1c06af34647db2c3c607cfceb7ee637622e3f22"} Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.368077 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wdnnh" event={"ID":"4da518e4-d3e3-437c-baa8-4fdef073dde5","Type":"ContainerStarted","Data":"3967987b7815929536a2df8bbd247c399a0c5e016ee2f89d5b12593431047e10"} Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.386059 4967 scope.go:117] "RemoveContainer" containerID="4f7d46ee97fd3c4286f45879fededea788052cf83ebeebcb539be452f2ae102d" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.426616 4967 scope.go:117] "RemoveContainer" containerID="27fb2b1225e386dcd7fecaaa0aa54ba1fdfafabfde6c3824ae07315313067542" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.458234 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-bs8w9"] Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.465292 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-bs8w9"] Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.466588 4967 scope.go:117] "RemoveContainer" containerID="9c9fa0559eb35ba67e11e36069ac8d0b2fec76399788eef668a08ad4353df3ab" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.493100 4967 scope.go:117] "RemoveContainer" containerID="d3312263e12bd42fbed0476ab5ace66b5497f62c9477865e130d67e3922e05cb" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.510310 4967 scope.go:117] "RemoveContainer" containerID="b03b83ddff429dbfa8fe052901aaa86a39ff0dd75477ffa8a6b6f4198f5819ca" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.530588 4967 scope.go:117] "RemoveContainer" containerID="4535cffb8c9d64e662e282ecb44f6817b4604866f0bb75015f3b888d5554f821" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.545515 4967 scope.go:117] "RemoveContainer" containerID="6e69dc603055aa221d44d11c049ddcc50cec568979911a310ed773a90341a499" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.562134 4967 scope.go:117] "RemoveContainer" containerID="115f7a33f3dedae8f59d2f45f465f95f48b99c824e895ef7a0be9fe7a1b61d23" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.577477 4967 scope.go:117] "RemoveContainer" containerID="94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.606036 4967 scope.go:117] "RemoveContainer" containerID="fab4c6588d3f2b44975eb2c248ad8b1cfef9b4b0fb797d0436123b63f9479df6" Oct 08 00:15:41 crc kubenswrapper[4967]: E1008 00:15:41.607007 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fab4c6588d3f2b44975eb2c248ad8b1cfef9b4b0fb797d0436123b63f9479df6\": container with ID starting with fab4c6588d3f2b44975eb2c248ad8b1cfef9b4b0fb797d0436123b63f9479df6 not found: ID does not exist" containerID="fab4c6588d3f2b44975eb2c248ad8b1cfef9b4b0fb797d0436123b63f9479df6" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.607058 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fab4c6588d3f2b44975eb2c248ad8b1cfef9b4b0fb797d0436123b63f9479df6"} err="failed to get container status \"fab4c6588d3f2b44975eb2c248ad8b1cfef9b4b0fb797d0436123b63f9479df6\": rpc error: code = NotFound desc = could not find container \"fab4c6588d3f2b44975eb2c248ad8b1cfef9b4b0fb797d0436123b63f9479df6\": container with ID starting with fab4c6588d3f2b44975eb2c248ad8b1cfef9b4b0fb797d0436123b63f9479df6 not found: ID does not exist" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.607101 4967 scope.go:117] "RemoveContainer" containerID="4f7d46ee97fd3c4286f45879fededea788052cf83ebeebcb539be452f2ae102d" Oct 08 00:15:41 crc kubenswrapper[4967]: E1008 00:15:41.607419 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f7d46ee97fd3c4286f45879fededea788052cf83ebeebcb539be452f2ae102d\": container with ID starting with 4f7d46ee97fd3c4286f45879fededea788052cf83ebeebcb539be452f2ae102d not found: ID does not exist" containerID="4f7d46ee97fd3c4286f45879fededea788052cf83ebeebcb539be452f2ae102d" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.607455 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f7d46ee97fd3c4286f45879fededea788052cf83ebeebcb539be452f2ae102d"} err="failed to get container status \"4f7d46ee97fd3c4286f45879fededea788052cf83ebeebcb539be452f2ae102d\": rpc error: code = NotFound desc = could not find container \"4f7d46ee97fd3c4286f45879fededea788052cf83ebeebcb539be452f2ae102d\": container with ID starting with 4f7d46ee97fd3c4286f45879fededea788052cf83ebeebcb539be452f2ae102d not found: ID does not exist" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.607480 4967 scope.go:117] "RemoveContainer" containerID="27fb2b1225e386dcd7fecaaa0aa54ba1fdfafabfde6c3824ae07315313067542" Oct 08 00:15:41 crc kubenswrapper[4967]: E1008 00:15:41.607850 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"27fb2b1225e386dcd7fecaaa0aa54ba1fdfafabfde6c3824ae07315313067542\": container with ID starting with 27fb2b1225e386dcd7fecaaa0aa54ba1fdfafabfde6c3824ae07315313067542 not found: ID does not exist" containerID="27fb2b1225e386dcd7fecaaa0aa54ba1fdfafabfde6c3824ae07315313067542" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.607888 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27fb2b1225e386dcd7fecaaa0aa54ba1fdfafabfde6c3824ae07315313067542"} err="failed to get container status \"27fb2b1225e386dcd7fecaaa0aa54ba1fdfafabfde6c3824ae07315313067542\": rpc error: code = NotFound desc = could not find container \"27fb2b1225e386dcd7fecaaa0aa54ba1fdfafabfde6c3824ae07315313067542\": container with ID starting with 27fb2b1225e386dcd7fecaaa0aa54ba1fdfafabfde6c3824ae07315313067542 not found: ID does not exist" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.607909 4967 scope.go:117] "RemoveContainer" containerID="9c9fa0559eb35ba67e11e36069ac8d0b2fec76399788eef668a08ad4353df3ab" Oct 08 00:15:41 crc kubenswrapper[4967]: E1008 00:15:41.608133 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c9fa0559eb35ba67e11e36069ac8d0b2fec76399788eef668a08ad4353df3ab\": container with ID starting with 9c9fa0559eb35ba67e11e36069ac8d0b2fec76399788eef668a08ad4353df3ab not found: ID does not exist" containerID="9c9fa0559eb35ba67e11e36069ac8d0b2fec76399788eef668a08ad4353df3ab" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.608166 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c9fa0559eb35ba67e11e36069ac8d0b2fec76399788eef668a08ad4353df3ab"} err="failed to get container status \"9c9fa0559eb35ba67e11e36069ac8d0b2fec76399788eef668a08ad4353df3ab\": rpc error: code = NotFound desc = could not find container \"9c9fa0559eb35ba67e11e36069ac8d0b2fec76399788eef668a08ad4353df3ab\": container with ID starting with 9c9fa0559eb35ba67e11e36069ac8d0b2fec76399788eef668a08ad4353df3ab not found: ID does not exist" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.608186 4967 scope.go:117] "RemoveContainer" containerID="d3312263e12bd42fbed0476ab5ace66b5497f62c9477865e130d67e3922e05cb" Oct 08 00:15:41 crc kubenswrapper[4967]: E1008 00:15:41.608458 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d3312263e12bd42fbed0476ab5ace66b5497f62c9477865e130d67e3922e05cb\": container with ID starting with d3312263e12bd42fbed0476ab5ace66b5497f62c9477865e130d67e3922e05cb not found: ID does not exist" containerID="d3312263e12bd42fbed0476ab5ace66b5497f62c9477865e130d67e3922e05cb" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.608477 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d3312263e12bd42fbed0476ab5ace66b5497f62c9477865e130d67e3922e05cb"} err="failed to get container status \"d3312263e12bd42fbed0476ab5ace66b5497f62c9477865e130d67e3922e05cb\": rpc error: code = NotFound desc = could not find container \"d3312263e12bd42fbed0476ab5ace66b5497f62c9477865e130d67e3922e05cb\": container with ID starting with d3312263e12bd42fbed0476ab5ace66b5497f62c9477865e130d67e3922e05cb not found: ID does not exist" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.608490 4967 scope.go:117] "RemoveContainer" containerID="b03b83ddff429dbfa8fe052901aaa86a39ff0dd75477ffa8a6b6f4198f5819ca" Oct 08 00:15:41 crc kubenswrapper[4967]: E1008 00:15:41.608898 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b03b83ddff429dbfa8fe052901aaa86a39ff0dd75477ffa8a6b6f4198f5819ca\": container with ID starting with b03b83ddff429dbfa8fe052901aaa86a39ff0dd75477ffa8a6b6f4198f5819ca not found: ID does not exist" containerID="b03b83ddff429dbfa8fe052901aaa86a39ff0dd75477ffa8a6b6f4198f5819ca" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.608925 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b03b83ddff429dbfa8fe052901aaa86a39ff0dd75477ffa8a6b6f4198f5819ca"} err="failed to get container status \"b03b83ddff429dbfa8fe052901aaa86a39ff0dd75477ffa8a6b6f4198f5819ca\": rpc error: code = NotFound desc = could not find container \"b03b83ddff429dbfa8fe052901aaa86a39ff0dd75477ffa8a6b6f4198f5819ca\": container with ID starting with b03b83ddff429dbfa8fe052901aaa86a39ff0dd75477ffa8a6b6f4198f5819ca not found: ID does not exist" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.608942 4967 scope.go:117] "RemoveContainer" containerID="4535cffb8c9d64e662e282ecb44f6817b4604866f0bb75015f3b888d5554f821" Oct 08 00:15:41 crc kubenswrapper[4967]: E1008 00:15:41.609322 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4535cffb8c9d64e662e282ecb44f6817b4604866f0bb75015f3b888d5554f821\": container with ID starting with 4535cffb8c9d64e662e282ecb44f6817b4604866f0bb75015f3b888d5554f821 not found: ID does not exist" containerID="4535cffb8c9d64e662e282ecb44f6817b4604866f0bb75015f3b888d5554f821" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.609352 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4535cffb8c9d64e662e282ecb44f6817b4604866f0bb75015f3b888d5554f821"} err="failed to get container status \"4535cffb8c9d64e662e282ecb44f6817b4604866f0bb75015f3b888d5554f821\": rpc error: code = NotFound desc = could not find container \"4535cffb8c9d64e662e282ecb44f6817b4604866f0bb75015f3b888d5554f821\": container with ID starting with 4535cffb8c9d64e662e282ecb44f6817b4604866f0bb75015f3b888d5554f821 not found: ID does not exist" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.609372 4967 scope.go:117] "RemoveContainer" containerID="6e69dc603055aa221d44d11c049ddcc50cec568979911a310ed773a90341a499" Oct 08 00:15:41 crc kubenswrapper[4967]: E1008 00:15:41.609709 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6e69dc603055aa221d44d11c049ddcc50cec568979911a310ed773a90341a499\": container with ID starting with 6e69dc603055aa221d44d11c049ddcc50cec568979911a310ed773a90341a499 not found: ID does not exist" containerID="6e69dc603055aa221d44d11c049ddcc50cec568979911a310ed773a90341a499" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.609733 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e69dc603055aa221d44d11c049ddcc50cec568979911a310ed773a90341a499"} err="failed to get container status \"6e69dc603055aa221d44d11c049ddcc50cec568979911a310ed773a90341a499\": rpc error: code = NotFound desc = could not find container \"6e69dc603055aa221d44d11c049ddcc50cec568979911a310ed773a90341a499\": container with ID starting with 6e69dc603055aa221d44d11c049ddcc50cec568979911a310ed773a90341a499 not found: ID does not exist" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.609746 4967 scope.go:117] "RemoveContainer" containerID="115f7a33f3dedae8f59d2f45f465f95f48b99c824e895ef7a0be9fe7a1b61d23" Oct 08 00:15:41 crc kubenswrapper[4967]: E1008 00:15:41.610087 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"115f7a33f3dedae8f59d2f45f465f95f48b99c824e895ef7a0be9fe7a1b61d23\": container with ID starting with 115f7a33f3dedae8f59d2f45f465f95f48b99c824e895ef7a0be9fe7a1b61d23 not found: ID does not exist" containerID="115f7a33f3dedae8f59d2f45f465f95f48b99c824e895ef7a0be9fe7a1b61d23" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.610115 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"115f7a33f3dedae8f59d2f45f465f95f48b99c824e895ef7a0be9fe7a1b61d23"} err="failed to get container status \"115f7a33f3dedae8f59d2f45f465f95f48b99c824e895ef7a0be9fe7a1b61d23\": rpc error: code = NotFound desc = could not find container \"115f7a33f3dedae8f59d2f45f465f95f48b99c824e895ef7a0be9fe7a1b61d23\": container with ID starting with 115f7a33f3dedae8f59d2f45f465f95f48b99c824e895ef7a0be9fe7a1b61d23 not found: ID does not exist" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.610136 4967 scope.go:117] "RemoveContainer" containerID="94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427" Oct 08 00:15:41 crc kubenswrapper[4967]: E1008 00:15:41.610382 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427\": container with ID starting with 94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427 not found: ID does not exist" containerID="94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.610404 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427"} err="failed to get container status \"94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427\": rpc error: code = NotFound desc = could not find container \"94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427\": container with ID starting with 94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427 not found: ID does not exist" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.610417 4967 scope.go:117] "RemoveContainer" containerID="fab4c6588d3f2b44975eb2c248ad8b1cfef9b4b0fb797d0436123b63f9479df6" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.610617 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fab4c6588d3f2b44975eb2c248ad8b1cfef9b4b0fb797d0436123b63f9479df6"} err="failed to get container status \"fab4c6588d3f2b44975eb2c248ad8b1cfef9b4b0fb797d0436123b63f9479df6\": rpc error: code = NotFound desc = could not find container \"fab4c6588d3f2b44975eb2c248ad8b1cfef9b4b0fb797d0436123b63f9479df6\": container with ID starting with fab4c6588d3f2b44975eb2c248ad8b1cfef9b4b0fb797d0436123b63f9479df6 not found: ID does not exist" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.610643 4967 scope.go:117] "RemoveContainer" containerID="4f7d46ee97fd3c4286f45879fededea788052cf83ebeebcb539be452f2ae102d" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.611133 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f7d46ee97fd3c4286f45879fededea788052cf83ebeebcb539be452f2ae102d"} err="failed to get container status \"4f7d46ee97fd3c4286f45879fededea788052cf83ebeebcb539be452f2ae102d\": rpc error: code = NotFound desc = could not find container \"4f7d46ee97fd3c4286f45879fededea788052cf83ebeebcb539be452f2ae102d\": container with ID starting with 4f7d46ee97fd3c4286f45879fededea788052cf83ebeebcb539be452f2ae102d not found: ID does not exist" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.611150 4967 scope.go:117] "RemoveContainer" containerID="27fb2b1225e386dcd7fecaaa0aa54ba1fdfafabfde6c3824ae07315313067542" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.611480 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27fb2b1225e386dcd7fecaaa0aa54ba1fdfafabfde6c3824ae07315313067542"} err="failed to get container status \"27fb2b1225e386dcd7fecaaa0aa54ba1fdfafabfde6c3824ae07315313067542\": rpc error: code = NotFound desc = could not find container \"27fb2b1225e386dcd7fecaaa0aa54ba1fdfafabfde6c3824ae07315313067542\": container with ID starting with 27fb2b1225e386dcd7fecaaa0aa54ba1fdfafabfde6c3824ae07315313067542 not found: ID does not exist" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.611496 4967 scope.go:117] "RemoveContainer" containerID="9c9fa0559eb35ba67e11e36069ac8d0b2fec76399788eef668a08ad4353df3ab" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.611838 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c9fa0559eb35ba67e11e36069ac8d0b2fec76399788eef668a08ad4353df3ab"} err="failed to get container status \"9c9fa0559eb35ba67e11e36069ac8d0b2fec76399788eef668a08ad4353df3ab\": rpc error: code = NotFound desc = could not find container \"9c9fa0559eb35ba67e11e36069ac8d0b2fec76399788eef668a08ad4353df3ab\": container with ID starting with 9c9fa0559eb35ba67e11e36069ac8d0b2fec76399788eef668a08ad4353df3ab not found: ID does not exist" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.611871 4967 scope.go:117] "RemoveContainer" containerID="d3312263e12bd42fbed0476ab5ace66b5497f62c9477865e130d67e3922e05cb" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.612132 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d3312263e12bd42fbed0476ab5ace66b5497f62c9477865e130d67e3922e05cb"} err="failed to get container status \"d3312263e12bd42fbed0476ab5ace66b5497f62c9477865e130d67e3922e05cb\": rpc error: code = NotFound desc = could not find container \"d3312263e12bd42fbed0476ab5ace66b5497f62c9477865e130d67e3922e05cb\": container with ID starting with d3312263e12bd42fbed0476ab5ace66b5497f62c9477865e130d67e3922e05cb not found: ID does not exist" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.612157 4967 scope.go:117] "RemoveContainer" containerID="b03b83ddff429dbfa8fe052901aaa86a39ff0dd75477ffa8a6b6f4198f5819ca" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.612407 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b03b83ddff429dbfa8fe052901aaa86a39ff0dd75477ffa8a6b6f4198f5819ca"} err="failed to get container status \"b03b83ddff429dbfa8fe052901aaa86a39ff0dd75477ffa8a6b6f4198f5819ca\": rpc error: code = NotFound desc = could not find container \"b03b83ddff429dbfa8fe052901aaa86a39ff0dd75477ffa8a6b6f4198f5819ca\": container with ID starting with b03b83ddff429dbfa8fe052901aaa86a39ff0dd75477ffa8a6b6f4198f5819ca not found: ID does not exist" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.612426 4967 scope.go:117] "RemoveContainer" containerID="4535cffb8c9d64e662e282ecb44f6817b4604866f0bb75015f3b888d5554f821" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.612744 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4535cffb8c9d64e662e282ecb44f6817b4604866f0bb75015f3b888d5554f821"} err="failed to get container status \"4535cffb8c9d64e662e282ecb44f6817b4604866f0bb75015f3b888d5554f821\": rpc error: code = NotFound desc = could not find container \"4535cffb8c9d64e662e282ecb44f6817b4604866f0bb75015f3b888d5554f821\": container with ID starting with 4535cffb8c9d64e662e282ecb44f6817b4604866f0bb75015f3b888d5554f821 not found: ID does not exist" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.612760 4967 scope.go:117] "RemoveContainer" containerID="6e69dc603055aa221d44d11c049ddcc50cec568979911a310ed773a90341a499" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.613072 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e69dc603055aa221d44d11c049ddcc50cec568979911a310ed773a90341a499"} err="failed to get container status \"6e69dc603055aa221d44d11c049ddcc50cec568979911a310ed773a90341a499\": rpc error: code = NotFound desc = could not find container \"6e69dc603055aa221d44d11c049ddcc50cec568979911a310ed773a90341a499\": container with ID starting with 6e69dc603055aa221d44d11c049ddcc50cec568979911a310ed773a90341a499 not found: ID does not exist" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.613089 4967 scope.go:117] "RemoveContainer" containerID="115f7a33f3dedae8f59d2f45f465f95f48b99c824e895ef7a0be9fe7a1b61d23" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.613347 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"115f7a33f3dedae8f59d2f45f465f95f48b99c824e895ef7a0be9fe7a1b61d23"} err="failed to get container status \"115f7a33f3dedae8f59d2f45f465f95f48b99c824e895ef7a0be9fe7a1b61d23\": rpc error: code = NotFound desc = could not find container \"115f7a33f3dedae8f59d2f45f465f95f48b99c824e895ef7a0be9fe7a1b61d23\": container with ID starting with 115f7a33f3dedae8f59d2f45f465f95f48b99c824e895ef7a0be9fe7a1b61d23 not found: ID does not exist" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.613364 4967 scope.go:117] "RemoveContainer" containerID="94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.613683 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427"} err="failed to get container status \"94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427\": rpc error: code = NotFound desc = could not find container \"94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427\": container with ID starting with 94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427 not found: ID does not exist" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.613707 4967 scope.go:117] "RemoveContainer" containerID="fab4c6588d3f2b44975eb2c248ad8b1cfef9b4b0fb797d0436123b63f9479df6" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.613966 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fab4c6588d3f2b44975eb2c248ad8b1cfef9b4b0fb797d0436123b63f9479df6"} err="failed to get container status \"fab4c6588d3f2b44975eb2c248ad8b1cfef9b4b0fb797d0436123b63f9479df6\": rpc error: code = NotFound desc = could not find container \"fab4c6588d3f2b44975eb2c248ad8b1cfef9b4b0fb797d0436123b63f9479df6\": container with ID starting with fab4c6588d3f2b44975eb2c248ad8b1cfef9b4b0fb797d0436123b63f9479df6 not found: ID does not exist" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.613988 4967 scope.go:117] "RemoveContainer" containerID="4f7d46ee97fd3c4286f45879fededea788052cf83ebeebcb539be452f2ae102d" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.614263 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f7d46ee97fd3c4286f45879fededea788052cf83ebeebcb539be452f2ae102d"} err="failed to get container status \"4f7d46ee97fd3c4286f45879fededea788052cf83ebeebcb539be452f2ae102d\": rpc error: code = NotFound desc = could not find container \"4f7d46ee97fd3c4286f45879fededea788052cf83ebeebcb539be452f2ae102d\": container with ID starting with 4f7d46ee97fd3c4286f45879fededea788052cf83ebeebcb539be452f2ae102d not found: ID does not exist" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.614278 4967 scope.go:117] "RemoveContainer" containerID="27fb2b1225e386dcd7fecaaa0aa54ba1fdfafabfde6c3824ae07315313067542" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.614608 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27fb2b1225e386dcd7fecaaa0aa54ba1fdfafabfde6c3824ae07315313067542"} err="failed to get container status \"27fb2b1225e386dcd7fecaaa0aa54ba1fdfafabfde6c3824ae07315313067542\": rpc error: code = NotFound desc = could not find container \"27fb2b1225e386dcd7fecaaa0aa54ba1fdfafabfde6c3824ae07315313067542\": container with ID starting with 27fb2b1225e386dcd7fecaaa0aa54ba1fdfafabfde6c3824ae07315313067542 not found: ID does not exist" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.614629 4967 scope.go:117] "RemoveContainer" containerID="9c9fa0559eb35ba67e11e36069ac8d0b2fec76399788eef668a08ad4353df3ab" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.614864 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c9fa0559eb35ba67e11e36069ac8d0b2fec76399788eef668a08ad4353df3ab"} err="failed to get container status \"9c9fa0559eb35ba67e11e36069ac8d0b2fec76399788eef668a08ad4353df3ab\": rpc error: code = NotFound desc = could not find container \"9c9fa0559eb35ba67e11e36069ac8d0b2fec76399788eef668a08ad4353df3ab\": container with ID starting with 9c9fa0559eb35ba67e11e36069ac8d0b2fec76399788eef668a08ad4353df3ab not found: ID does not exist" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.614887 4967 scope.go:117] "RemoveContainer" containerID="d3312263e12bd42fbed0476ab5ace66b5497f62c9477865e130d67e3922e05cb" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.615272 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d3312263e12bd42fbed0476ab5ace66b5497f62c9477865e130d67e3922e05cb"} err="failed to get container status \"d3312263e12bd42fbed0476ab5ace66b5497f62c9477865e130d67e3922e05cb\": rpc error: code = NotFound desc = could not find container \"d3312263e12bd42fbed0476ab5ace66b5497f62c9477865e130d67e3922e05cb\": container with ID starting with d3312263e12bd42fbed0476ab5ace66b5497f62c9477865e130d67e3922e05cb not found: ID does not exist" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.615292 4967 scope.go:117] "RemoveContainer" containerID="b03b83ddff429dbfa8fe052901aaa86a39ff0dd75477ffa8a6b6f4198f5819ca" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.615541 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b03b83ddff429dbfa8fe052901aaa86a39ff0dd75477ffa8a6b6f4198f5819ca"} err="failed to get container status \"b03b83ddff429dbfa8fe052901aaa86a39ff0dd75477ffa8a6b6f4198f5819ca\": rpc error: code = NotFound desc = could not find container \"b03b83ddff429dbfa8fe052901aaa86a39ff0dd75477ffa8a6b6f4198f5819ca\": container with ID starting with b03b83ddff429dbfa8fe052901aaa86a39ff0dd75477ffa8a6b6f4198f5819ca not found: ID does not exist" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.615563 4967 scope.go:117] "RemoveContainer" containerID="4535cffb8c9d64e662e282ecb44f6817b4604866f0bb75015f3b888d5554f821" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.615857 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4535cffb8c9d64e662e282ecb44f6817b4604866f0bb75015f3b888d5554f821"} err="failed to get container status \"4535cffb8c9d64e662e282ecb44f6817b4604866f0bb75015f3b888d5554f821\": rpc error: code = NotFound desc = could not find container \"4535cffb8c9d64e662e282ecb44f6817b4604866f0bb75015f3b888d5554f821\": container with ID starting with 4535cffb8c9d64e662e282ecb44f6817b4604866f0bb75015f3b888d5554f821 not found: ID does not exist" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.615878 4967 scope.go:117] "RemoveContainer" containerID="6e69dc603055aa221d44d11c049ddcc50cec568979911a310ed773a90341a499" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.616117 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e69dc603055aa221d44d11c049ddcc50cec568979911a310ed773a90341a499"} err="failed to get container status \"6e69dc603055aa221d44d11c049ddcc50cec568979911a310ed773a90341a499\": rpc error: code = NotFound desc = could not find container \"6e69dc603055aa221d44d11c049ddcc50cec568979911a310ed773a90341a499\": container with ID starting with 6e69dc603055aa221d44d11c049ddcc50cec568979911a310ed773a90341a499 not found: ID does not exist" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.616142 4967 scope.go:117] "RemoveContainer" containerID="115f7a33f3dedae8f59d2f45f465f95f48b99c824e895ef7a0be9fe7a1b61d23" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.616476 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"115f7a33f3dedae8f59d2f45f465f95f48b99c824e895ef7a0be9fe7a1b61d23"} err="failed to get container status \"115f7a33f3dedae8f59d2f45f465f95f48b99c824e895ef7a0be9fe7a1b61d23\": rpc error: code = NotFound desc = could not find container \"115f7a33f3dedae8f59d2f45f465f95f48b99c824e895ef7a0be9fe7a1b61d23\": container with ID starting with 115f7a33f3dedae8f59d2f45f465f95f48b99c824e895ef7a0be9fe7a1b61d23 not found: ID does not exist" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.616536 4967 scope.go:117] "RemoveContainer" containerID="94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.616913 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427"} err="failed to get container status \"94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427\": rpc error: code = NotFound desc = could not find container \"94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427\": container with ID starting with 94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427 not found: ID does not exist" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.616939 4967 scope.go:117] "RemoveContainer" containerID="fab4c6588d3f2b44975eb2c248ad8b1cfef9b4b0fb797d0436123b63f9479df6" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.617358 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fab4c6588d3f2b44975eb2c248ad8b1cfef9b4b0fb797d0436123b63f9479df6"} err="failed to get container status \"fab4c6588d3f2b44975eb2c248ad8b1cfef9b4b0fb797d0436123b63f9479df6\": rpc error: code = NotFound desc = could not find container \"fab4c6588d3f2b44975eb2c248ad8b1cfef9b4b0fb797d0436123b63f9479df6\": container with ID starting with fab4c6588d3f2b44975eb2c248ad8b1cfef9b4b0fb797d0436123b63f9479df6 not found: ID does not exist" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.617387 4967 scope.go:117] "RemoveContainer" containerID="4f7d46ee97fd3c4286f45879fededea788052cf83ebeebcb539be452f2ae102d" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.617635 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f7d46ee97fd3c4286f45879fededea788052cf83ebeebcb539be452f2ae102d"} err="failed to get container status \"4f7d46ee97fd3c4286f45879fededea788052cf83ebeebcb539be452f2ae102d\": rpc error: code = NotFound desc = could not find container \"4f7d46ee97fd3c4286f45879fededea788052cf83ebeebcb539be452f2ae102d\": container with ID starting with 4f7d46ee97fd3c4286f45879fededea788052cf83ebeebcb539be452f2ae102d not found: ID does not exist" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.617657 4967 scope.go:117] "RemoveContainer" containerID="27fb2b1225e386dcd7fecaaa0aa54ba1fdfafabfde6c3824ae07315313067542" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.618002 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27fb2b1225e386dcd7fecaaa0aa54ba1fdfafabfde6c3824ae07315313067542"} err="failed to get container status \"27fb2b1225e386dcd7fecaaa0aa54ba1fdfafabfde6c3824ae07315313067542\": rpc error: code = NotFound desc = could not find container \"27fb2b1225e386dcd7fecaaa0aa54ba1fdfafabfde6c3824ae07315313067542\": container with ID starting with 27fb2b1225e386dcd7fecaaa0aa54ba1fdfafabfde6c3824ae07315313067542 not found: ID does not exist" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.618026 4967 scope.go:117] "RemoveContainer" containerID="9c9fa0559eb35ba67e11e36069ac8d0b2fec76399788eef668a08ad4353df3ab" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.618347 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c9fa0559eb35ba67e11e36069ac8d0b2fec76399788eef668a08ad4353df3ab"} err="failed to get container status \"9c9fa0559eb35ba67e11e36069ac8d0b2fec76399788eef668a08ad4353df3ab\": rpc error: code = NotFound desc = could not find container \"9c9fa0559eb35ba67e11e36069ac8d0b2fec76399788eef668a08ad4353df3ab\": container with ID starting with 9c9fa0559eb35ba67e11e36069ac8d0b2fec76399788eef668a08ad4353df3ab not found: ID does not exist" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.618370 4967 scope.go:117] "RemoveContainer" containerID="d3312263e12bd42fbed0476ab5ace66b5497f62c9477865e130d67e3922e05cb" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.618632 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d3312263e12bd42fbed0476ab5ace66b5497f62c9477865e130d67e3922e05cb"} err="failed to get container status \"d3312263e12bd42fbed0476ab5ace66b5497f62c9477865e130d67e3922e05cb\": rpc error: code = NotFound desc = could not find container \"d3312263e12bd42fbed0476ab5ace66b5497f62c9477865e130d67e3922e05cb\": container with ID starting with d3312263e12bd42fbed0476ab5ace66b5497f62c9477865e130d67e3922e05cb not found: ID does not exist" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.618657 4967 scope.go:117] "RemoveContainer" containerID="b03b83ddff429dbfa8fe052901aaa86a39ff0dd75477ffa8a6b6f4198f5819ca" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.619007 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b03b83ddff429dbfa8fe052901aaa86a39ff0dd75477ffa8a6b6f4198f5819ca"} err="failed to get container status \"b03b83ddff429dbfa8fe052901aaa86a39ff0dd75477ffa8a6b6f4198f5819ca\": rpc error: code = NotFound desc = could not find container \"b03b83ddff429dbfa8fe052901aaa86a39ff0dd75477ffa8a6b6f4198f5819ca\": container with ID starting with b03b83ddff429dbfa8fe052901aaa86a39ff0dd75477ffa8a6b6f4198f5819ca not found: ID does not exist" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.619027 4967 scope.go:117] "RemoveContainer" containerID="4535cffb8c9d64e662e282ecb44f6817b4604866f0bb75015f3b888d5554f821" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.619274 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4535cffb8c9d64e662e282ecb44f6817b4604866f0bb75015f3b888d5554f821"} err="failed to get container status \"4535cffb8c9d64e662e282ecb44f6817b4604866f0bb75015f3b888d5554f821\": rpc error: code = NotFound desc = could not find container \"4535cffb8c9d64e662e282ecb44f6817b4604866f0bb75015f3b888d5554f821\": container with ID starting with 4535cffb8c9d64e662e282ecb44f6817b4604866f0bb75015f3b888d5554f821 not found: ID does not exist" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.619292 4967 scope.go:117] "RemoveContainer" containerID="6e69dc603055aa221d44d11c049ddcc50cec568979911a310ed773a90341a499" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.619554 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e69dc603055aa221d44d11c049ddcc50cec568979911a310ed773a90341a499"} err="failed to get container status \"6e69dc603055aa221d44d11c049ddcc50cec568979911a310ed773a90341a499\": rpc error: code = NotFound desc = could not find container \"6e69dc603055aa221d44d11c049ddcc50cec568979911a310ed773a90341a499\": container with ID starting with 6e69dc603055aa221d44d11c049ddcc50cec568979911a310ed773a90341a499 not found: ID does not exist" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.619572 4967 scope.go:117] "RemoveContainer" containerID="115f7a33f3dedae8f59d2f45f465f95f48b99c824e895ef7a0be9fe7a1b61d23" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.619828 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"115f7a33f3dedae8f59d2f45f465f95f48b99c824e895ef7a0be9fe7a1b61d23"} err="failed to get container status \"115f7a33f3dedae8f59d2f45f465f95f48b99c824e895ef7a0be9fe7a1b61d23\": rpc error: code = NotFound desc = could not find container \"115f7a33f3dedae8f59d2f45f465f95f48b99c824e895ef7a0be9fe7a1b61d23\": container with ID starting with 115f7a33f3dedae8f59d2f45f465f95f48b99c824e895ef7a0be9fe7a1b61d23 not found: ID does not exist" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.619856 4967 scope.go:117] "RemoveContainer" containerID="94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.620211 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427"} err="failed to get container status \"94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427\": rpc error: code = NotFound desc = could not find container \"94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427\": container with ID starting with 94ac8a782d810f5bb613fb702f29157da8599a51a33c070f0e8204c7d9698427 not found: ID does not exist" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.620233 4967 scope.go:117] "RemoveContainer" containerID="fab4c6588d3f2b44975eb2c248ad8b1cfef9b4b0fb797d0436123b63f9479df6" Oct 08 00:15:41 crc kubenswrapper[4967]: I1008 00:15:41.620637 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fab4c6588d3f2b44975eb2c248ad8b1cfef9b4b0fb797d0436123b63f9479df6"} err="failed to get container status \"fab4c6588d3f2b44975eb2c248ad8b1cfef9b4b0fb797d0436123b63f9479df6\": rpc error: code = NotFound desc = could not find container \"fab4c6588d3f2b44975eb2c248ad8b1cfef9b4b0fb797d0436123b63f9479df6\": container with ID starting with fab4c6588d3f2b44975eb2c248ad8b1cfef9b4b0fb797d0436123b63f9479df6 not found: ID does not exist" Oct 08 00:15:42 crc kubenswrapper[4967]: I1008 00:15:42.132402 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4f1af2ba-8907-4d14-afa1-ffde207c80f4" path="/var/lib/kubelet/pods/4f1af2ba-8907-4d14-afa1-ffde207c80f4/volumes" Oct 08 00:15:42 crc kubenswrapper[4967]: I1008 00:15:42.291640 4967 scope.go:117] "RemoveContainer" containerID="5ae6a42d62117bc7aeca9ebb2c6aa58dc5c18e8f8bbabcbb287f458de8c03e89" Oct 08 00:15:42 crc kubenswrapper[4967]: I1008 00:15:42.313108 4967 scope.go:117] "RemoveContainer" containerID="e2f7ac3d2d6f7a8672da3676f8bc923f89da5b022af990fe7f3d8be5b95a2743" Oct 08 00:15:42 crc kubenswrapper[4967]: I1008 00:15:42.336611 4967 scope.go:117] "RemoveContainer" containerID="e7db1c6944d8b20adde1074eead4d289d796a9369b0a8ed3736e49af93449098" Oct 08 00:15:42 crc kubenswrapper[4967]: I1008 00:15:42.355083 4967 scope.go:117] "RemoveContainer" containerID="82dfe6c420d804b2d7daef4b5cfe3f87b77b369287c4d65553fa93f5f24f589d" Oct 08 00:15:42 crc kubenswrapper[4967]: I1008 00:15:42.382789 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wdnnh" event={"ID":"4da518e4-d3e3-437c-baa8-4fdef073dde5","Type":"ContainerStarted","Data":"8f3be054749ae51213ee61dc2e805b80cb9a77fd7da5b25fe28c23edb0d1c182"} Oct 08 00:15:42 crc kubenswrapper[4967]: I1008 00:15:42.382868 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wdnnh" event={"ID":"4da518e4-d3e3-437c-baa8-4fdef073dde5","Type":"ContainerStarted","Data":"27e4060a25ea74663bb249220c258ca164d1a891e39a6c985def30c2a8376f41"} Oct 08 00:15:42 crc kubenswrapper[4967]: I1008 00:15:42.382898 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wdnnh" event={"ID":"4da518e4-d3e3-437c-baa8-4fdef073dde5","Type":"ContainerStarted","Data":"54ccbc8f0243ff226e66f2c1f073143bbbe3ef896dcd4e9ee2a3b8ff28211b86"} Oct 08 00:15:42 crc kubenswrapper[4967]: I1008 00:15:42.382931 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wdnnh" event={"ID":"4da518e4-d3e3-437c-baa8-4fdef073dde5","Type":"ContainerStarted","Data":"ab18843751b5e61dc1de4e8e126737d53307a58fe2571585f3cd8eaa1c284104"} Oct 08 00:15:42 crc kubenswrapper[4967]: I1008 00:15:42.382957 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wdnnh" event={"ID":"4da518e4-d3e3-437c-baa8-4fdef073dde5","Type":"ContainerStarted","Data":"76140e95c1c879d932ac656a2f1a2625dbfa7cd8a3ac7687908767514efe9e99"} Oct 08 00:15:42 crc kubenswrapper[4967]: I1008 00:15:42.382981 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wdnnh" event={"ID":"4da518e4-d3e3-437c-baa8-4fdef073dde5","Type":"ContainerStarted","Data":"3fc84efb9b8b198356777069d2cd0bf5ed565f08d1a07ba12c914f994579f5d4"} Oct 08 00:15:43 crc kubenswrapper[4967]: I1008 00:15:43.391911 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-npz2s_c71501b1-aa9e-420e-b1bd-7e15365e7adc/kube-multus/1.log" Oct 08 00:15:44 crc kubenswrapper[4967]: I1008 00:15:44.407275 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wdnnh" event={"ID":"4da518e4-d3e3-437c-baa8-4fdef073dde5","Type":"ContainerStarted","Data":"13a211ca8644c37f0115c3e6149d451ae1ea98274f3db650193654b140a0f23c"} Oct 08 00:15:47 crc kubenswrapper[4967]: I1008 00:15:47.429993 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wdnnh" event={"ID":"4da518e4-d3e3-437c-baa8-4fdef073dde5","Type":"ContainerStarted","Data":"569996ecad46f5f03ecebf56033f789eec1a77be5b5cb5b637636724eecee042"} Oct 08 00:15:47 crc kubenswrapper[4967]: I1008 00:15:47.431625 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-wdnnh" Oct 08 00:15:47 crc kubenswrapper[4967]: I1008 00:15:47.431642 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-wdnnh" Oct 08 00:15:47 crc kubenswrapper[4967]: I1008 00:15:47.431651 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-wdnnh" Oct 08 00:15:47 crc kubenswrapper[4967]: I1008 00:15:47.458351 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-wdnnh" Oct 08 00:15:47 crc kubenswrapper[4967]: I1008 00:15:47.459935 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-wdnnh" podStartSLOduration=7.459920543 podStartE2EDuration="7.459920543s" podCreationTimestamp="2025-10-08 00:15:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:15:47.458493442 +0000 UTC m=+546.109028111" watchObservedRunningTime="2025-10-08 00:15:47.459920543 +0000 UTC m=+546.110455192" Oct 08 00:15:47 crc kubenswrapper[4967]: I1008 00:15:47.467185 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-wdnnh" Oct 08 00:15:53 crc kubenswrapper[4967]: I1008 00:15:53.117832 4967 scope.go:117] "RemoveContainer" containerID="9010a5d6165a64f986f9101dbbbcaf8a7d3fa5af85c14be4bf3aa8ed8536ad4e" Oct 08 00:15:53 crc kubenswrapper[4967]: I1008 00:15:53.480105 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-npz2s_c71501b1-aa9e-420e-b1bd-7e15365e7adc/kube-multus/1.log" Oct 08 00:15:53 crc kubenswrapper[4967]: I1008 00:15:53.480469 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-npz2s" event={"ID":"c71501b1-aa9e-420e-b1bd-7e15365e7adc","Type":"ContainerStarted","Data":"09a0a67e4bf86172551881f956d330a31f341d9a247ab39460be5a788543adc6"} Oct 08 00:16:06 crc kubenswrapper[4967]: I1008 00:16:06.921034 4967 patch_prober.go:28] interesting pod/machine-config-daemon-xvdtp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 00:16:06 crc kubenswrapper[4967]: I1008 00:16:06.921529 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" podUID="9461162a-c27d-465e-9ca7-2d3936877cd8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 00:16:11 crc kubenswrapper[4967]: I1008 00:16:11.138516 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-wdnnh" Oct 08 00:16:36 crc kubenswrapper[4967]: I1008 00:16:36.919915 4967 patch_prober.go:28] interesting pod/machine-config-daemon-xvdtp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 00:16:36 crc kubenswrapper[4967]: I1008 00:16:36.920630 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" podUID="9461162a-c27d-465e-9ca7-2d3936877cd8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 00:16:43 crc kubenswrapper[4967]: I1008 00:16:43.780353 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-pjrvj"] Oct 08 00:16:43 crc kubenswrapper[4967]: I1008 00:16:43.781747 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-pjrvj" podUID="09cce4f0-143f-44b5-8343-56409fa4d0a3" containerName="registry-server" containerID="cri-o://a474d3f13903ad7e8d8340d2c559e1b6d672592b2d1e0aaf231f717dfafb3e9b" gracePeriod=30 Oct 08 00:16:44 crc kubenswrapper[4967]: I1008 00:16:44.197838 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pjrvj" Oct 08 00:16:44 crc kubenswrapper[4967]: I1008 00:16:44.285270 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8k5m8\" (UniqueName: \"kubernetes.io/projected/09cce4f0-143f-44b5-8343-56409fa4d0a3-kube-api-access-8k5m8\") pod \"09cce4f0-143f-44b5-8343-56409fa4d0a3\" (UID: \"09cce4f0-143f-44b5-8343-56409fa4d0a3\") " Oct 08 00:16:44 crc kubenswrapper[4967]: I1008 00:16:44.285388 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09cce4f0-143f-44b5-8343-56409fa4d0a3-utilities\") pod \"09cce4f0-143f-44b5-8343-56409fa4d0a3\" (UID: \"09cce4f0-143f-44b5-8343-56409fa4d0a3\") " Oct 08 00:16:44 crc kubenswrapper[4967]: I1008 00:16:44.285467 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09cce4f0-143f-44b5-8343-56409fa4d0a3-catalog-content\") pod \"09cce4f0-143f-44b5-8343-56409fa4d0a3\" (UID: \"09cce4f0-143f-44b5-8343-56409fa4d0a3\") " Oct 08 00:16:44 crc kubenswrapper[4967]: I1008 00:16:44.287458 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/09cce4f0-143f-44b5-8343-56409fa4d0a3-utilities" (OuterVolumeSpecName: "utilities") pod "09cce4f0-143f-44b5-8343-56409fa4d0a3" (UID: "09cce4f0-143f-44b5-8343-56409fa4d0a3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:16:44 crc kubenswrapper[4967]: I1008 00:16:44.293298 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09cce4f0-143f-44b5-8343-56409fa4d0a3-kube-api-access-8k5m8" (OuterVolumeSpecName: "kube-api-access-8k5m8") pod "09cce4f0-143f-44b5-8343-56409fa4d0a3" (UID: "09cce4f0-143f-44b5-8343-56409fa4d0a3"). InnerVolumeSpecName "kube-api-access-8k5m8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:16:44 crc kubenswrapper[4967]: I1008 00:16:44.309751 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/09cce4f0-143f-44b5-8343-56409fa4d0a3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "09cce4f0-143f-44b5-8343-56409fa4d0a3" (UID: "09cce4f0-143f-44b5-8343-56409fa4d0a3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:16:44 crc kubenswrapper[4967]: I1008 00:16:44.387158 4967 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09cce4f0-143f-44b5-8343-56409fa4d0a3-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 00:16:44 crc kubenswrapper[4967]: I1008 00:16:44.387248 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8k5m8\" (UniqueName: \"kubernetes.io/projected/09cce4f0-143f-44b5-8343-56409fa4d0a3-kube-api-access-8k5m8\") on node \"crc\" DevicePath \"\"" Oct 08 00:16:44 crc kubenswrapper[4967]: I1008 00:16:44.387272 4967 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09cce4f0-143f-44b5-8343-56409fa4d0a3-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 00:16:44 crc kubenswrapper[4967]: I1008 00:16:44.815998 4967 generic.go:334] "Generic (PLEG): container finished" podID="09cce4f0-143f-44b5-8343-56409fa4d0a3" containerID="a474d3f13903ad7e8d8340d2c559e1b6d672592b2d1e0aaf231f717dfafb3e9b" exitCode=0 Oct 08 00:16:44 crc kubenswrapper[4967]: I1008 00:16:44.816034 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pjrvj" event={"ID":"09cce4f0-143f-44b5-8343-56409fa4d0a3","Type":"ContainerDied","Data":"a474d3f13903ad7e8d8340d2c559e1b6d672592b2d1e0aaf231f717dfafb3e9b"} Oct 08 00:16:44 crc kubenswrapper[4967]: I1008 00:16:44.816066 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pjrvj" Oct 08 00:16:44 crc kubenswrapper[4967]: I1008 00:16:44.816948 4967 scope.go:117] "RemoveContainer" containerID="a474d3f13903ad7e8d8340d2c559e1b6d672592b2d1e0aaf231f717dfafb3e9b" Oct 08 00:16:44 crc kubenswrapper[4967]: I1008 00:16:44.816881 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pjrvj" event={"ID":"09cce4f0-143f-44b5-8343-56409fa4d0a3","Type":"ContainerDied","Data":"e56e99111d4ae2c4e8449a95c1f5a52f0b3923e23568ac7460e2f235ede7eab1"} Oct 08 00:16:44 crc kubenswrapper[4967]: I1008 00:16:44.834087 4967 scope.go:117] "RemoveContainer" containerID="2cff2052c829abc3002ad4560c04638e0e16ee835fea82c4782b0ebbdc8e99b6" Oct 08 00:16:44 crc kubenswrapper[4967]: I1008 00:16:44.840930 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-pjrvj"] Oct 08 00:16:44 crc kubenswrapper[4967]: I1008 00:16:44.845607 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-pjrvj"] Oct 08 00:16:44 crc kubenswrapper[4967]: I1008 00:16:44.873328 4967 scope.go:117] "RemoveContainer" containerID="8ef952affb669ec11c952345092a78a93b0107f73ef03af4901e9a30007da121" Oct 08 00:16:44 crc kubenswrapper[4967]: I1008 00:16:44.888655 4967 scope.go:117] "RemoveContainer" containerID="a474d3f13903ad7e8d8340d2c559e1b6d672592b2d1e0aaf231f717dfafb3e9b" Oct 08 00:16:44 crc kubenswrapper[4967]: E1008 00:16:44.889534 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a474d3f13903ad7e8d8340d2c559e1b6d672592b2d1e0aaf231f717dfafb3e9b\": container with ID starting with a474d3f13903ad7e8d8340d2c559e1b6d672592b2d1e0aaf231f717dfafb3e9b not found: ID does not exist" containerID="a474d3f13903ad7e8d8340d2c559e1b6d672592b2d1e0aaf231f717dfafb3e9b" Oct 08 00:16:44 crc kubenswrapper[4967]: I1008 00:16:44.889606 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a474d3f13903ad7e8d8340d2c559e1b6d672592b2d1e0aaf231f717dfafb3e9b"} err="failed to get container status \"a474d3f13903ad7e8d8340d2c559e1b6d672592b2d1e0aaf231f717dfafb3e9b\": rpc error: code = NotFound desc = could not find container \"a474d3f13903ad7e8d8340d2c559e1b6d672592b2d1e0aaf231f717dfafb3e9b\": container with ID starting with a474d3f13903ad7e8d8340d2c559e1b6d672592b2d1e0aaf231f717dfafb3e9b not found: ID does not exist" Oct 08 00:16:44 crc kubenswrapper[4967]: I1008 00:16:44.889626 4967 scope.go:117] "RemoveContainer" containerID="2cff2052c829abc3002ad4560c04638e0e16ee835fea82c4782b0ebbdc8e99b6" Oct 08 00:16:44 crc kubenswrapper[4967]: E1008 00:16:44.890115 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2cff2052c829abc3002ad4560c04638e0e16ee835fea82c4782b0ebbdc8e99b6\": container with ID starting with 2cff2052c829abc3002ad4560c04638e0e16ee835fea82c4782b0ebbdc8e99b6 not found: ID does not exist" containerID="2cff2052c829abc3002ad4560c04638e0e16ee835fea82c4782b0ebbdc8e99b6" Oct 08 00:16:44 crc kubenswrapper[4967]: I1008 00:16:44.890171 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2cff2052c829abc3002ad4560c04638e0e16ee835fea82c4782b0ebbdc8e99b6"} err="failed to get container status \"2cff2052c829abc3002ad4560c04638e0e16ee835fea82c4782b0ebbdc8e99b6\": rpc error: code = NotFound desc = could not find container \"2cff2052c829abc3002ad4560c04638e0e16ee835fea82c4782b0ebbdc8e99b6\": container with ID starting with 2cff2052c829abc3002ad4560c04638e0e16ee835fea82c4782b0ebbdc8e99b6 not found: ID does not exist" Oct 08 00:16:44 crc kubenswrapper[4967]: I1008 00:16:44.890206 4967 scope.go:117] "RemoveContainer" containerID="8ef952affb669ec11c952345092a78a93b0107f73ef03af4901e9a30007da121" Oct 08 00:16:44 crc kubenswrapper[4967]: E1008 00:16:44.890500 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8ef952affb669ec11c952345092a78a93b0107f73ef03af4901e9a30007da121\": container with ID starting with 8ef952affb669ec11c952345092a78a93b0107f73ef03af4901e9a30007da121 not found: ID does not exist" containerID="8ef952affb669ec11c952345092a78a93b0107f73ef03af4901e9a30007da121" Oct 08 00:16:44 crc kubenswrapper[4967]: I1008 00:16:44.890543 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8ef952affb669ec11c952345092a78a93b0107f73ef03af4901e9a30007da121"} err="failed to get container status \"8ef952affb669ec11c952345092a78a93b0107f73ef03af4901e9a30007da121\": rpc error: code = NotFound desc = could not find container \"8ef952affb669ec11c952345092a78a93b0107f73ef03af4901e9a30007da121\": container with ID starting with 8ef952affb669ec11c952345092a78a93b0107f73ef03af4901e9a30007da121 not found: ID does not exist" Oct 08 00:16:46 crc kubenswrapper[4967]: I1008 00:16:46.125107 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09cce4f0-143f-44b5-8343-56409fa4d0a3" path="/var/lib/kubelet/pods/09cce4f0-143f-44b5-8343-56409fa4d0a3/volumes" Oct 08 00:16:47 crc kubenswrapper[4967]: I1008 00:16:47.317936 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d4n2bh"] Oct 08 00:16:47 crc kubenswrapper[4967]: E1008 00:16:47.318217 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09cce4f0-143f-44b5-8343-56409fa4d0a3" containerName="extract-utilities" Oct 08 00:16:47 crc kubenswrapper[4967]: I1008 00:16:47.318239 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="09cce4f0-143f-44b5-8343-56409fa4d0a3" containerName="extract-utilities" Oct 08 00:16:47 crc kubenswrapper[4967]: E1008 00:16:47.318264 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09cce4f0-143f-44b5-8343-56409fa4d0a3" containerName="extract-content" Oct 08 00:16:47 crc kubenswrapper[4967]: I1008 00:16:47.318275 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="09cce4f0-143f-44b5-8343-56409fa4d0a3" containerName="extract-content" Oct 08 00:16:47 crc kubenswrapper[4967]: E1008 00:16:47.318287 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09cce4f0-143f-44b5-8343-56409fa4d0a3" containerName="registry-server" Oct 08 00:16:47 crc kubenswrapper[4967]: I1008 00:16:47.318300 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="09cce4f0-143f-44b5-8343-56409fa4d0a3" containerName="registry-server" Oct 08 00:16:47 crc kubenswrapper[4967]: I1008 00:16:47.318450 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="09cce4f0-143f-44b5-8343-56409fa4d0a3" containerName="registry-server" Oct 08 00:16:47 crc kubenswrapper[4967]: I1008 00:16:47.319512 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d4n2bh" Oct 08 00:16:47 crc kubenswrapper[4967]: I1008 00:16:47.321850 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 08 00:16:47 crc kubenswrapper[4967]: I1008 00:16:47.346690 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d4n2bh"] Oct 08 00:16:47 crc kubenswrapper[4967]: I1008 00:16:47.425487 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kjbwl\" (UniqueName: \"kubernetes.io/projected/55341c94-8af6-4a61-8fa8-913ced0e2732-kube-api-access-kjbwl\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d4n2bh\" (UID: \"55341c94-8af6-4a61-8fa8-913ced0e2732\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d4n2bh" Oct 08 00:16:47 crc kubenswrapper[4967]: I1008 00:16:47.425579 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/55341c94-8af6-4a61-8fa8-913ced0e2732-bundle\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d4n2bh\" (UID: \"55341c94-8af6-4a61-8fa8-913ced0e2732\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d4n2bh" Oct 08 00:16:47 crc kubenswrapper[4967]: I1008 00:16:47.425659 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/55341c94-8af6-4a61-8fa8-913ced0e2732-util\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d4n2bh\" (UID: \"55341c94-8af6-4a61-8fa8-913ced0e2732\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d4n2bh" Oct 08 00:16:47 crc kubenswrapper[4967]: I1008 00:16:47.526626 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kjbwl\" (UniqueName: \"kubernetes.io/projected/55341c94-8af6-4a61-8fa8-913ced0e2732-kube-api-access-kjbwl\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d4n2bh\" (UID: \"55341c94-8af6-4a61-8fa8-913ced0e2732\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d4n2bh" Oct 08 00:16:47 crc kubenswrapper[4967]: I1008 00:16:47.526714 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/55341c94-8af6-4a61-8fa8-913ced0e2732-bundle\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d4n2bh\" (UID: \"55341c94-8af6-4a61-8fa8-913ced0e2732\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d4n2bh" Oct 08 00:16:47 crc kubenswrapper[4967]: I1008 00:16:47.526760 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/55341c94-8af6-4a61-8fa8-913ced0e2732-util\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d4n2bh\" (UID: \"55341c94-8af6-4a61-8fa8-913ced0e2732\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d4n2bh" Oct 08 00:16:47 crc kubenswrapper[4967]: I1008 00:16:47.527250 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/55341c94-8af6-4a61-8fa8-913ced0e2732-util\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d4n2bh\" (UID: \"55341c94-8af6-4a61-8fa8-913ced0e2732\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d4n2bh" Oct 08 00:16:47 crc kubenswrapper[4967]: I1008 00:16:47.527413 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/55341c94-8af6-4a61-8fa8-913ced0e2732-bundle\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d4n2bh\" (UID: \"55341c94-8af6-4a61-8fa8-913ced0e2732\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d4n2bh" Oct 08 00:16:47 crc kubenswrapper[4967]: I1008 00:16:47.554274 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kjbwl\" (UniqueName: \"kubernetes.io/projected/55341c94-8af6-4a61-8fa8-913ced0e2732-kube-api-access-kjbwl\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d4n2bh\" (UID: \"55341c94-8af6-4a61-8fa8-913ced0e2732\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d4n2bh" Oct 08 00:16:47 crc kubenswrapper[4967]: I1008 00:16:47.700549 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d4n2bh" Oct 08 00:16:48 crc kubenswrapper[4967]: I1008 00:16:48.124299 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d4n2bh"] Oct 08 00:16:48 crc kubenswrapper[4967]: I1008 00:16:48.846338 4967 generic.go:334] "Generic (PLEG): container finished" podID="55341c94-8af6-4a61-8fa8-913ced0e2732" containerID="3dfb4de9c56aa4f381147bb57232a10fc63db95cd4f37e3631f34d4146d3dfce" exitCode=0 Oct 08 00:16:48 crc kubenswrapper[4967]: I1008 00:16:48.846420 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d4n2bh" event={"ID":"55341c94-8af6-4a61-8fa8-913ced0e2732","Type":"ContainerDied","Data":"3dfb4de9c56aa4f381147bb57232a10fc63db95cd4f37e3631f34d4146d3dfce"} Oct 08 00:16:48 crc kubenswrapper[4967]: I1008 00:16:48.846459 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d4n2bh" event={"ID":"55341c94-8af6-4a61-8fa8-913ced0e2732","Type":"ContainerStarted","Data":"bec69af38a28e804d3c919b738672d88c079b5cf941de7b942444a8644a66687"} Oct 08 00:16:48 crc kubenswrapper[4967]: I1008 00:16:48.848633 4967 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 00:16:54 crc kubenswrapper[4967]: I1008 00:16:54.519260 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fd24kq"] Oct 08 00:16:54 crc kubenswrapper[4967]: I1008 00:16:54.521662 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fd24kq" Oct 08 00:16:54 crc kubenswrapper[4967]: I1008 00:16:54.533055 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fd24kq"] Oct 08 00:16:54 crc kubenswrapper[4967]: I1008 00:16:54.625339 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w8skj\" (UniqueName: \"kubernetes.io/projected/3221812d-48b9-4361-a75e-3472b715d548-kube-api-access-w8skj\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fd24kq\" (UID: \"3221812d-48b9-4361-a75e-3472b715d548\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fd24kq" Oct 08 00:16:54 crc kubenswrapper[4967]: I1008 00:16:54.625501 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3221812d-48b9-4361-a75e-3472b715d548-util\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fd24kq\" (UID: \"3221812d-48b9-4361-a75e-3472b715d548\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fd24kq" Oct 08 00:16:54 crc kubenswrapper[4967]: I1008 00:16:54.625563 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3221812d-48b9-4361-a75e-3472b715d548-bundle\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fd24kq\" (UID: \"3221812d-48b9-4361-a75e-3472b715d548\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fd24kq" Oct 08 00:16:54 crc kubenswrapper[4967]: I1008 00:16:54.727326 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3221812d-48b9-4361-a75e-3472b715d548-bundle\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fd24kq\" (UID: \"3221812d-48b9-4361-a75e-3472b715d548\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fd24kq" Oct 08 00:16:54 crc kubenswrapper[4967]: I1008 00:16:54.727474 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w8skj\" (UniqueName: \"kubernetes.io/projected/3221812d-48b9-4361-a75e-3472b715d548-kube-api-access-w8skj\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fd24kq\" (UID: \"3221812d-48b9-4361-a75e-3472b715d548\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fd24kq" Oct 08 00:16:54 crc kubenswrapper[4967]: I1008 00:16:54.727633 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3221812d-48b9-4361-a75e-3472b715d548-util\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fd24kq\" (UID: \"3221812d-48b9-4361-a75e-3472b715d548\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fd24kq" Oct 08 00:16:54 crc kubenswrapper[4967]: I1008 00:16:54.728901 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3221812d-48b9-4361-a75e-3472b715d548-util\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fd24kq\" (UID: \"3221812d-48b9-4361-a75e-3472b715d548\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fd24kq" Oct 08 00:16:54 crc kubenswrapper[4967]: I1008 00:16:54.729171 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3221812d-48b9-4361-a75e-3472b715d548-bundle\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fd24kq\" (UID: \"3221812d-48b9-4361-a75e-3472b715d548\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fd24kq" Oct 08 00:16:54 crc kubenswrapper[4967]: I1008 00:16:54.761319 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w8skj\" (UniqueName: \"kubernetes.io/projected/3221812d-48b9-4361-a75e-3472b715d548-kube-api-access-w8skj\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fd24kq\" (UID: \"3221812d-48b9-4361-a75e-3472b715d548\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fd24kq" Oct 08 00:16:54 crc kubenswrapper[4967]: I1008 00:16:54.839981 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fd24kq" Oct 08 00:16:55 crc kubenswrapper[4967]: I1008 00:16:55.052053 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fd24kq"] Oct 08 00:16:55 crc kubenswrapper[4967]: W1008 00:16:55.061496 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3221812d_48b9_4361_a75e_3472b715d548.slice/crio-86a87f6c3e22bcabb93d92333ac1ec723e20baad67da17cdce30c94a0ca43eba WatchSource:0}: Error finding container 86a87f6c3e22bcabb93d92333ac1ec723e20baad67da17cdce30c94a0ca43eba: Status 404 returned error can't find the container with id 86a87f6c3e22bcabb93d92333ac1ec723e20baad67da17cdce30c94a0ca43eba Oct 08 00:16:55 crc kubenswrapper[4967]: I1008 00:16:55.301251 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/dec6855936450d9745a6250315030408d50c8ea57f37cefa88be37c927pvlxl"] Oct 08 00:16:55 crc kubenswrapper[4967]: I1008 00:16:55.304948 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/dec6855936450d9745a6250315030408d50c8ea57f37cefa88be37c927pvlxl" Oct 08 00:16:55 crc kubenswrapper[4967]: I1008 00:16:55.314579 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/dec6855936450d9745a6250315030408d50c8ea57f37cefa88be37c927pvlxl"] Oct 08 00:16:55 crc kubenswrapper[4967]: I1008 00:16:55.437323 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/330d296a-13cf-4883-8d5b-af3896f592b8-util\") pod \"dec6855936450d9745a6250315030408d50c8ea57f37cefa88be37c927pvlxl\" (UID: \"330d296a-13cf-4883-8d5b-af3896f592b8\") " pod="openshift-marketplace/dec6855936450d9745a6250315030408d50c8ea57f37cefa88be37c927pvlxl" Oct 08 00:16:55 crc kubenswrapper[4967]: I1008 00:16:55.437726 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hrzvk\" (UniqueName: \"kubernetes.io/projected/330d296a-13cf-4883-8d5b-af3896f592b8-kube-api-access-hrzvk\") pod \"dec6855936450d9745a6250315030408d50c8ea57f37cefa88be37c927pvlxl\" (UID: \"330d296a-13cf-4883-8d5b-af3896f592b8\") " pod="openshift-marketplace/dec6855936450d9745a6250315030408d50c8ea57f37cefa88be37c927pvlxl" Oct 08 00:16:55 crc kubenswrapper[4967]: I1008 00:16:55.437899 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/330d296a-13cf-4883-8d5b-af3896f592b8-bundle\") pod \"dec6855936450d9745a6250315030408d50c8ea57f37cefa88be37c927pvlxl\" (UID: \"330d296a-13cf-4883-8d5b-af3896f592b8\") " pod="openshift-marketplace/dec6855936450d9745a6250315030408d50c8ea57f37cefa88be37c927pvlxl" Oct 08 00:16:55 crc kubenswrapper[4967]: I1008 00:16:55.539591 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hrzvk\" (UniqueName: \"kubernetes.io/projected/330d296a-13cf-4883-8d5b-af3896f592b8-kube-api-access-hrzvk\") pod \"dec6855936450d9745a6250315030408d50c8ea57f37cefa88be37c927pvlxl\" (UID: \"330d296a-13cf-4883-8d5b-af3896f592b8\") " pod="openshift-marketplace/dec6855936450d9745a6250315030408d50c8ea57f37cefa88be37c927pvlxl" Oct 08 00:16:55 crc kubenswrapper[4967]: I1008 00:16:55.539703 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/330d296a-13cf-4883-8d5b-af3896f592b8-bundle\") pod \"dec6855936450d9745a6250315030408d50c8ea57f37cefa88be37c927pvlxl\" (UID: \"330d296a-13cf-4883-8d5b-af3896f592b8\") " pod="openshift-marketplace/dec6855936450d9745a6250315030408d50c8ea57f37cefa88be37c927pvlxl" Oct 08 00:16:55 crc kubenswrapper[4967]: I1008 00:16:55.539763 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/330d296a-13cf-4883-8d5b-af3896f592b8-util\") pod \"dec6855936450d9745a6250315030408d50c8ea57f37cefa88be37c927pvlxl\" (UID: \"330d296a-13cf-4883-8d5b-af3896f592b8\") " pod="openshift-marketplace/dec6855936450d9745a6250315030408d50c8ea57f37cefa88be37c927pvlxl" Oct 08 00:16:55 crc kubenswrapper[4967]: I1008 00:16:55.540524 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/330d296a-13cf-4883-8d5b-af3896f592b8-util\") pod \"dec6855936450d9745a6250315030408d50c8ea57f37cefa88be37c927pvlxl\" (UID: \"330d296a-13cf-4883-8d5b-af3896f592b8\") " pod="openshift-marketplace/dec6855936450d9745a6250315030408d50c8ea57f37cefa88be37c927pvlxl" Oct 08 00:16:55 crc kubenswrapper[4967]: I1008 00:16:55.540713 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/330d296a-13cf-4883-8d5b-af3896f592b8-bundle\") pod \"dec6855936450d9745a6250315030408d50c8ea57f37cefa88be37c927pvlxl\" (UID: \"330d296a-13cf-4883-8d5b-af3896f592b8\") " pod="openshift-marketplace/dec6855936450d9745a6250315030408d50c8ea57f37cefa88be37c927pvlxl" Oct 08 00:16:55 crc kubenswrapper[4967]: I1008 00:16:55.574268 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hrzvk\" (UniqueName: \"kubernetes.io/projected/330d296a-13cf-4883-8d5b-af3896f592b8-kube-api-access-hrzvk\") pod \"dec6855936450d9745a6250315030408d50c8ea57f37cefa88be37c927pvlxl\" (UID: \"330d296a-13cf-4883-8d5b-af3896f592b8\") " pod="openshift-marketplace/dec6855936450d9745a6250315030408d50c8ea57f37cefa88be37c927pvlxl" Oct 08 00:16:55 crc kubenswrapper[4967]: I1008 00:16:55.677302 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/dec6855936450d9745a6250315030408d50c8ea57f37cefa88be37c927pvlxl" Oct 08 00:16:55 crc kubenswrapper[4967]: I1008 00:16:55.892519 4967 generic.go:334] "Generic (PLEG): container finished" podID="3221812d-48b9-4361-a75e-3472b715d548" containerID="cbff6449739001f638031710bbd60173952608edc88a33cfef03d9449881692c" exitCode=0 Oct 08 00:16:55 crc kubenswrapper[4967]: I1008 00:16:55.892582 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fd24kq" event={"ID":"3221812d-48b9-4361-a75e-3472b715d548","Type":"ContainerDied","Data":"cbff6449739001f638031710bbd60173952608edc88a33cfef03d9449881692c"} Oct 08 00:16:55 crc kubenswrapper[4967]: I1008 00:16:55.892829 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fd24kq" event={"ID":"3221812d-48b9-4361-a75e-3472b715d548","Type":"ContainerStarted","Data":"86a87f6c3e22bcabb93d92333ac1ec723e20baad67da17cdce30c94a0ca43eba"} Oct 08 00:16:56 crc kubenswrapper[4967]: I1008 00:16:56.107980 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/dec6855936450d9745a6250315030408d50c8ea57f37cefa88be37c927pvlxl"] Oct 08 00:16:56 crc kubenswrapper[4967]: W1008 00:16:56.124894 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod330d296a_13cf_4883_8d5b_af3896f592b8.slice/crio-11c48e71ed313178ae6361715122777337a9f464ceb4a5d58f72afb02a2ef385 WatchSource:0}: Error finding container 11c48e71ed313178ae6361715122777337a9f464ceb4a5d58f72afb02a2ef385: Status 404 returned error can't find the container with id 11c48e71ed313178ae6361715122777337a9f464ceb4a5d58f72afb02a2ef385 Oct 08 00:16:56 crc kubenswrapper[4967]: I1008 00:16:56.901041 4967 generic.go:334] "Generic (PLEG): container finished" podID="330d296a-13cf-4883-8d5b-af3896f592b8" containerID="cdeb723518fc7ca5c9878a5de816079bf3c9923b7e65bfe31641cb004d439120" exitCode=0 Oct 08 00:16:56 crc kubenswrapper[4967]: I1008 00:16:56.901136 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/dec6855936450d9745a6250315030408d50c8ea57f37cefa88be37c927pvlxl" event={"ID":"330d296a-13cf-4883-8d5b-af3896f592b8","Type":"ContainerDied","Data":"cdeb723518fc7ca5c9878a5de816079bf3c9923b7e65bfe31641cb004d439120"} Oct 08 00:16:56 crc kubenswrapper[4967]: I1008 00:16:56.901167 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/dec6855936450d9745a6250315030408d50c8ea57f37cefa88be37c927pvlxl" event={"ID":"330d296a-13cf-4883-8d5b-af3896f592b8","Type":"ContainerStarted","Data":"11c48e71ed313178ae6361715122777337a9f464ceb4a5d58f72afb02a2ef385"} Oct 08 00:16:56 crc kubenswrapper[4967]: I1008 00:16:56.908126 4967 generic.go:334] "Generic (PLEG): container finished" podID="55341c94-8af6-4a61-8fa8-913ced0e2732" containerID="f21f81e6cc00cf1a4f7af4bb4c1f6866e939f5804fc8bd70b2e20a4d9533a66e" exitCode=0 Oct 08 00:16:56 crc kubenswrapper[4967]: I1008 00:16:56.908179 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d4n2bh" event={"ID":"55341c94-8af6-4a61-8fa8-913ced0e2732","Type":"ContainerDied","Data":"f21f81e6cc00cf1a4f7af4bb4c1f6866e939f5804fc8bd70b2e20a4d9533a66e"} Oct 08 00:16:57 crc kubenswrapper[4967]: I1008 00:16:57.917126 4967 generic.go:334] "Generic (PLEG): container finished" podID="55341c94-8af6-4a61-8fa8-913ced0e2732" containerID="d7753445220fb69c0ebdf1264af6e8d8dcb2538a0c9ea42b2d3008b9461e2f4a" exitCode=0 Oct 08 00:16:57 crc kubenswrapper[4967]: I1008 00:16:57.917228 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d4n2bh" event={"ID":"55341c94-8af6-4a61-8fa8-913ced0e2732","Type":"ContainerDied","Data":"d7753445220fb69c0ebdf1264af6e8d8dcb2538a0c9ea42b2d3008b9461e2f4a"} Oct 08 00:16:58 crc kubenswrapper[4967]: I1008 00:16:58.929099 4967 generic.go:334] "Generic (PLEG): container finished" podID="330d296a-13cf-4883-8d5b-af3896f592b8" containerID="b7920b5fe6d7e51e41dc5e3225a376ab6df180b7902819832eb4b4f5274f96c6" exitCode=0 Oct 08 00:16:58 crc kubenswrapper[4967]: I1008 00:16:58.929196 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/dec6855936450d9745a6250315030408d50c8ea57f37cefa88be37c927pvlxl" event={"ID":"330d296a-13cf-4883-8d5b-af3896f592b8","Type":"ContainerDied","Data":"b7920b5fe6d7e51e41dc5e3225a376ab6df180b7902819832eb4b4f5274f96c6"} Oct 08 00:16:58 crc kubenswrapper[4967]: I1008 00:16:58.932904 4967 generic.go:334] "Generic (PLEG): container finished" podID="3221812d-48b9-4361-a75e-3472b715d548" containerID="aa4c920494893e12da4ac5d9c1e992dee35fc483b8d277434f24ff829770988b" exitCode=0 Oct 08 00:16:58 crc kubenswrapper[4967]: I1008 00:16:58.933207 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fd24kq" event={"ID":"3221812d-48b9-4361-a75e-3472b715d548","Type":"ContainerDied","Data":"aa4c920494893e12da4ac5d9c1e992dee35fc483b8d277434f24ff829770988b"} Oct 08 00:16:59 crc kubenswrapper[4967]: I1008 00:16:59.206740 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d4n2bh" Oct 08 00:16:59 crc kubenswrapper[4967]: I1008 00:16:59.291489 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kjbwl\" (UniqueName: \"kubernetes.io/projected/55341c94-8af6-4a61-8fa8-913ced0e2732-kube-api-access-kjbwl\") pod \"55341c94-8af6-4a61-8fa8-913ced0e2732\" (UID: \"55341c94-8af6-4a61-8fa8-913ced0e2732\") " Oct 08 00:16:59 crc kubenswrapper[4967]: I1008 00:16:59.291556 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/55341c94-8af6-4a61-8fa8-913ced0e2732-util\") pod \"55341c94-8af6-4a61-8fa8-913ced0e2732\" (UID: \"55341c94-8af6-4a61-8fa8-913ced0e2732\") " Oct 08 00:16:59 crc kubenswrapper[4967]: I1008 00:16:59.291599 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/55341c94-8af6-4a61-8fa8-913ced0e2732-bundle\") pod \"55341c94-8af6-4a61-8fa8-913ced0e2732\" (UID: \"55341c94-8af6-4a61-8fa8-913ced0e2732\") " Oct 08 00:16:59 crc kubenswrapper[4967]: I1008 00:16:59.294337 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/55341c94-8af6-4a61-8fa8-913ced0e2732-bundle" (OuterVolumeSpecName: "bundle") pod "55341c94-8af6-4a61-8fa8-913ced0e2732" (UID: "55341c94-8af6-4a61-8fa8-913ced0e2732"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:16:59 crc kubenswrapper[4967]: I1008 00:16:59.300569 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55341c94-8af6-4a61-8fa8-913ced0e2732-kube-api-access-kjbwl" (OuterVolumeSpecName: "kube-api-access-kjbwl") pod "55341c94-8af6-4a61-8fa8-913ced0e2732" (UID: "55341c94-8af6-4a61-8fa8-913ced0e2732"). InnerVolumeSpecName "kube-api-access-kjbwl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:16:59 crc kubenswrapper[4967]: I1008 00:16:59.311169 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/55341c94-8af6-4a61-8fa8-913ced0e2732-util" (OuterVolumeSpecName: "util") pod "55341c94-8af6-4a61-8fa8-913ced0e2732" (UID: "55341c94-8af6-4a61-8fa8-913ced0e2732"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:16:59 crc kubenswrapper[4967]: I1008 00:16:59.392807 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kjbwl\" (UniqueName: \"kubernetes.io/projected/55341c94-8af6-4a61-8fa8-913ced0e2732-kube-api-access-kjbwl\") on node \"crc\" DevicePath \"\"" Oct 08 00:16:59 crc kubenswrapper[4967]: I1008 00:16:59.392855 4967 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/55341c94-8af6-4a61-8fa8-913ced0e2732-util\") on node \"crc\" DevicePath \"\"" Oct 08 00:16:59 crc kubenswrapper[4967]: I1008 00:16:59.392872 4967 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/55341c94-8af6-4a61-8fa8-913ced0e2732-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 00:16:59 crc kubenswrapper[4967]: I1008 00:16:59.943798 4967 generic.go:334] "Generic (PLEG): container finished" podID="330d296a-13cf-4883-8d5b-af3896f592b8" containerID="4e7dba8ffd03f709c0927dbcfb00fe4ebc806c53a88ff3bc25df939a02d2354a" exitCode=0 Oct 08 00:16:59 crc kubenswrapper[4967]: I1008 00:16:59.944162 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/dec6855936450d9745a6250315030408d50c8ea57f37cefa88be37c927pvlxl" event={"ID":"330d296a-13cf-4883-8d5b-af3896f592b8","Type":"ContainerDied","Data":"4e7dba8ffd03f709c0927dbcfb00fe4ebc806c53a88ff3bc25df939a02d2354a"} Oct 08 00:16:59 crc kubenswrapper[4967]: I1008 00:16:59.949077 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d4n2bh" event={"ID":"55341c94-8af6-4a61-8fa8-913ced0e2732","Type":"ContainerDied","Data":"bec69af38a28e804d3c919b738672d88c079b5cf941de7b942444a8644a66687"} Oct 08 00:16:59 crc kubenswrapper[4967]: I1008 00:16:59.949131 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bec69af38a28e804d3c919b738672d88c079b5cf941de7b942444a8644a66687" Oct 08 00:16:59 crc kubenswrapper[4967]: I1008 00:16:59.949206 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d4n2bh" Oct 08 00:16:59 crc kubenswrapper[4967]: I1008 00:16:59.954706 4967 generic.go:334] "Generic (PLEG): container finished" podID="3221812d-48b9-4361-a75e-3472b715d548" containerID="43aceb1fc6f6d98583f3a75d3864d15ac32b690eb08bda153f25ef0a034a5302" exitCode=0 Oct 08 00:16:59 crc kubenswrapper[4967]: I1008 00:16:59.954775 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fd24kq" event={"ID":"3221812d-48b9-4361-a75e-3472b715d548","Type":"ContainerDied","Data":"43aceb1fc6f6d98583f3a75d3864d15ac32b690eb08bda153f25ef0a034a5302"} Oct 08 00:17:00 crc kubenswrapper[4967]: I1008 00:17:00.720535 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69dw25g"] Oct 08 00:17:00 crc kubenswrapper[4967]: E1008 00:17:00.720860 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55341c94-8af6-4a61-8fa8-913ced0e2732" containerName="extract" Oct 08 00:17:00 crc kubenswrapper[4967]: I1008 00:17:00.720888 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="55341c94-8af6-4a61-8fa8-913ced0e2732" containerName="extract" Oct 08 00:17:00 crc kubenswrapper[4967]: E1008 00:17:00.720908 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55341c94-8af6-4a61-8fa8-913ced0e2732" containerName="util" Oct 08 00:17:00 crc kubenswrapper[4967]: I1008 00:17:00.720921 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="55341c94-8af6-4a61-8fa8-913ced0e2732" containerName="util" Oct 08 00:17:00 crc kubenswrapper[4967]: E1008 00:17:00.720960 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55341c94-8af6-4a61-8fa8-913ced0e2732" containerName="pull" Oct 08 00:17:00 crc kubenswrapper[4967]: I1008 00:17:00.720971 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="55341c94-8af6-4a61-8fa8-913ced0e2732" containerName="pull" Oct 08 00:17:00 crc kubenswrapper[4967]: I1008 00:17:00.721157 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="55341c94-8af6-4a61-8fa8-913ced0e2732" containerName="extract" Oct 08 00:17:00 crc kubenswrapper[4967]: I1008 00:17:00.722371 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69dw25g" Oct 08 00:17:00 crc kubenswrapper[4967]: I1008 00:17:00.747065 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69dw25g"] Oct 08 00:17:00 crc kubenswrapper[4967]: I1008 00:17:00.813188 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/277b8c1b-6e5d-44f2-b17b-594c3b367dab-bundle\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69dw25g\" (UID: \"277b8c1b-6e5d-44f2-b17b-594c3b367dab\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69dw25g" Oct 08 00:17:00 crc kubenswrapper[4967]: I1008 00:17:00.813244 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/277b8c1b-6e5d-44f2-b17b-594c3b367dab-util\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69dw25g\" (UID: \"277b8c1b-6e5d-44f2-b17b-594c3b367dab\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69dw25g" Oct 08 00:17:00 crc kubenswrapper[4967]: I1008 00:17:00.813383 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wvnc9\" (UniqueName: \"kubernetes.io/projected/277b8c1b-6e5d-44f2-b17b-594c3b367dab-kube-api-access-wvnc9\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69dw25g\" (UID: \"277b8c1b-6e5d-44f2-b17b-594c3b367dab\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69dw25g" Oct 08 00:17:00 crc kubenswrapper[4967]: I1008 00:17:00.914567 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/277b8c1b-6e5d-44f2-b17b-594c3b367dab-bundle\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69dw25g\" (UID: \"277b8c1b-6e5d-44f2-b17b-594c3b367dab\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69dw25g" Oct 08 00:17:00 crc kubenswrapper[4967]: I1008 00:17:00.914631 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/277b8c1b-6e5d-44f2-b17b-594c3b367dab-util\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69dw25g\" (UID: \"277b8c1b-6e5d-44f2-b17b-594c3b367dab\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69dw25g" Oct 08 00:17:00 crc kubenswrapper[4967]: I1008 00:17:00.914721 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wvnc9\" (UniqueName: \"kubernetes.io/projected/277b8c1b-6e5d-44f2-b17b-594c3b367dab-kube-api-access-wvnc9\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69dw25g\" (UID: \"277b8c1b-6e5d-44f2-b17b-594c3b367dab\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69dw25g" Oct 08 00:17:00 crc kubenswrapper[4967]: I1008 00:17:00.915473 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/277b8c1b-6e5d-44f2-b17b-594c3b367dab-bundle\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69dw25g\" (UID: \"277b8c1b-6e5d-44f2-b17b-594c3b367dab\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69dw25g" Oct 08 00:17:00 crc kubenswrapper[4967]: I1008 00:17:00.915542 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/277b8c1b-6e5d-44f2-b17b-594c3b367dab-util\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69dw25g\" (UID: \"277b8c1b-6e5d-44f2-b17b-594c3b367dab\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69dw25g" Oct 08 00:17:00 crc kubenswrapper[4967]: I1008 00:17:00.949357 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wvnc9\" (UniqueName: \"kubernetes.io/projected/277b8c1b-6e5d-44f2-b17b-594c3b367dab-kube-api-access-wvnc9\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69dw25g\" (UID: \"277b8c1b-6e5d-44f2-b17b-594c3b367dab\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69dw25g" Oct 08 00:17:01 crc kubenswrapper[4967]: I1008 00:17:01.049759 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69dw25g" Oct 08 00:17:01 crc kubenswrapper[4967]: I1008 00:17:01.294819 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/dec6855936450d9745a6250315030408d50c8ea57f37cefa88be37c927pvlxl" Oct 08 00:17:01 crc kubenswrapper[4967]: I1008 00:17:01.296992 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fd24kq" Oct 08 00:17:01 crc kubenswrapper[4967]: I1008 00:17:01.352651 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69dw25g"] Oct 08 00:17:01 crc kubenswrapper[4967]: W1008 00:17:01.356503 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod277b8c1b_6e5d_44f2_b17b_594c3b367dab.slice/crio-0c880f7673f8747141a82aeb557efba58ac20b3b64b7dbbe98e4db4a52ef8546 WatchSource:0}: Error finding container 0c880f7673f8747141a82aeb557efba58ac20b3b64b7dbbe98e4db4a52ef8546: Status 404 returned error can't find the container with id 0c880f7673f8747141a82aeb557efba58ac20b3b64b7dbbe98e4db4a52ef8546 Oct 08 00:17:01 crc kubenswrapper[4967]: I1008 00:17:01.420460 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3221812d-48b9-4361-a75e-3472b715d548-bundle\") pod \"3221812d-48b9-4361-a75e-3472b715d548\" (UID: \"3221812d-48b9-4361-a75e-3472b715d548\") " Oct 08 00:17:01 crc kubenswrapper[4967]: I1008 00:17:01.420691 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/330d296a-13cf-4883-8d5b-af3896f592b8-bundle\") pod \"330d296a-13cf-4883-8d5b-af3896f592b8\" (UID: \"330d296a-13cf-4883-8d5b-af3896f592b8\") " Oct 08 00:17:01 crc kubenswrapper[4967]: I1008 00:17:01.420762 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/330d296a-13cf-4883-8d5b-af3896f592b8-util\") pod \"330d296a-13cf-4883-8d5b-af3896f592b8\" (UID: \"330d296a-13cf-4883-8d5b-af3896f592b8\") " Oct 08 00:17:01 crc kubenswrapper[4967]: I1008 00:17:01.420790 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3221812d-48b9-4361-a75e-3472b715d548-util\") pod \"3221812d-48b9-4361-a75e-3472b715d548\" (UID: \"3221812d-48b9-4361-a75e-3472b715d548\") " Oct 08 00:17:01 crc kubenswrapper[4967]: I1008 00:17:01.420810 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hrzvk\" (UniqueName: \"kubernetes.io/projected/330d296a-13cf-4883-8d5b-af3896f592b8-kube-api-access-hrzvk\") pod \"330d296a-13cf-4883-8d5b-af3896f592b8\" (UID: \"330d296a-13cf-4883-8d5b-af3896f592b8\") " Oct 08 00:17:01 crc kubenswrapper[4967]: I1008 00:17:01.420849 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w8skj\" (UniqueName: \"kubernetes.io/projected/3221812d-48b9-4361-a75e-3472b715d548-kube-api-access-w8skj\") pod \"3221812d-48b9-4361-a75e-3472b715d548\" (UID: \"3221812d-48b9-4361-a75e-3472b715d548\") " Oct 08 00:17:01 crc kubenswrapper[4967]: I1008 00:17:01.421437 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3221812d-48b9-4361-a75e-3472b715d548-bundle" (OuterVolumeSpecName: "bundle") pod "3221812d-48b9-4361-a75e-3472b715d548" (UID: "3221812d-48b9-4361-a75e-3472b715d548"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:17:01 crc kubenswrapper[4967]: I1008 00:17:01.422190 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/330d296a-13cf-4883-8d5b-af3896f592b8-bundle" (OuterVolumeSpecName: "bundle") pod "330d296a-13cf-4883-8d5b-af3896f592b8" (UID: "330d296a-13cf-4883-8d5b-af3896f592b8"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:17:01 crc kubenswrapper[4967]: I1008 00:17:01.426533 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/330d296a-13cf-4883-8d5b-af3896f592b8-kube-api-access-hrzvk" (OuterVolumeSpecName: "kube-api-access-hrzvk") pod "330d296a-13cf-4883-8d5b-af3896f592b8" (UID: "330d296a-13cf-4883-8d5b-af3896f592b8"). InnerVolumeSpecName "kube-api-access-hrzvk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:17:01 crc kubenswrapper[4967]: I1008 00:17:01.426636 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3221812d-48b9-4361-a75e-3472b715d548-kube-api-access-w8skj" (OuterVolumeSpecName: "kube-api-access-w8skj") pod "3221812d-48b9-4361-a75e-3472b715d548" (UID: "3221812d-48b9-4361-a75e-3472b715d548"). InnerVolumeSpecName "kube-api-access-w8skj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:17:01 crc kubenswrapper[4967]: I1008 00:17:01.434602 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/330d296a-13cf-4883-8d5b-af3896f592b8-util" (OuterVolumeSpecName: "util") pod "330d296a-13cf-4883-8d5b-af3896f592b8" (UID: "330d296a-13cf-4883-8d5b-af3896f592b8"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:17:01 crc kubenswrapper[4967]: I1008 00:17:01.436958 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3221812d-48b9-4361-a75e-3472b715d548-util" (OuterVolumeSpecName: "util") pod "3221812d-48b9-4361-a75e-3472b715d548" (UID: "3221812d-48b9-4361-a75e-3472b715d548"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:17:01 crc kubenswrapper[4967]: I1008 00:17:01.522482 4967 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3221812d-48b9-4361-a75e-3472b715d548-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 00:17:01 crc kubenswrapper[4967]: I1008 00:17:01.522537 4967 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/330d296a-13cf-4883-8d5b-af3896f592b8-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 00:17:01 crc kubenswrapper[4967]: I1008 00:17:01.522559 4967 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/330d296a-13cf-4883-8d5b-af3896f592b8-util\") on node \"crc\" DevicePath \"\"" Oct 08 00:17:01 crc kubenswrapper[4967]: I1008 00:17:01.522575 4967 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3221812d-48b9-4361-a75e-3472b715d548-util\") on node \"crc\" DevicePath \"\"" Oct 08 00:17:01 crc kubenswrapper[4967]: I1008 00:17:01.522593 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hrzvk\" (UniqueName: \"kubernetes.io/projected/330d296a-13cf-4883-8d5b-af3896f592b8-kube-api-access-hrzvk\") on node \"crc\" DevicePath \"\"" Oct 08 00:17:01 crc kubenswrapper[4967]: I1008 00:17:01.522618 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w8skj\" (UniqueName: \"kubernetes.io/projected/3221812d-48b9-4361-a75e-3472b715d548-kube-api-access-w8skj\") on node \"crc\" DevicePath \"\"" Oct 08 00:17:01 crc kubenswrapper[4967]: I1008 00:17:01.979174 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fd24kq" event={"ID":"3221812d-48b9-4361-a75e-3472b715d548","Type":"ContainerDied","Data":"86a87f6c3e22bcabb93d92333ac1ec723e20baad67da17cdce30c94a0ca43eba"} Oct 08 00:17:01 crc kubenswrapper[4967]: I1008 00:17:01.979282 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="86a87f6c3e22bcabb93d92333ac1ec723e20baad67da17cdce30c94a0ca43eba" Oct 08 00:17:01 crc kubenswrapper[4967]: I1008 00:17:01.979202 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fd24kq" Oct 08 00:17:01 crc kubenswrapper[4967]: I1008 00:17:01.981576 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/dec6855936450d9745a6250315030408d50c8ea57f37cefa88be37c927pvlxl" event={"ID":"330d296a-13cf-4883-8d5b-af3896f592b8","Type":"ContainerDied","Data":"11c48e71ed313178ae6361715122777337a9f464ceb4a5d58f72afb02a2ef385"} Oct 08 00:17:01 crc kubenswrapper[4967]: I1008 00:17:01.981617 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/dec6855936450d9745a6250315030408d50c8ea57f37cefa88be37c927pvlxl" Oct 08 00:17:01 crc kubenswrapper[4967]: I1008 00:17:01.981623 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="11c48e71ed313178ae6361715122777337a9f464ceb4a5d58f72afb02a2ef385" Oct 08 00:17:01 crc kubenswrapper[4967]: I1008 00:17:01.984074 4967 generic.go:334] "Generic (PLEG): container finished" podID="277b8c1b-6e5d-44f2-b17b-594c3b367dab" containerID="76456129c694846495fbea9b5958cf6f36298b87edca9d10c386b91dcf0da0a7" exitCode=0 Oct 08 00:17:01 crc kubenswrapper[4967]: I1008 00:17:01.984119 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69dw25g" event={"ID":"277b8c1b-6e5d-44f2-b17b-594c3b367dab","Type":"ContainerDied","Data":"76456129c694846495fbea9b5958cf6f36298b87edca9d10c386b91dcf0da0a7"} Oct 08 00:17:01 crc kubenswrapper[4967]: I1008 00:17:01.984146 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69dw25g" event={"ID":"277b8c1b-6e5d-44f2-b17b-594c3b367dab","Type":"ContainerStarted","Data":"0c880f7673f8747141a82aeb557efba58ac20b3b64b7dbbe98e4db4a52ef8546"} Oct 08 00:17:06 crc kubenswrapper[4967]: I1008 00:17:06.017621 4967 generic.go:334] "Generic (PLEG): container finished" podID="277b8c1b-6e5d-44f2-b17b-594c3b367dab" containerID="6528f6595e6a3b8c89cedd72e92f2571f2d5bc726915cd1c08392ceedbc137b4" exitCode=0 Oct 08 00:17:06 crc kubenswrapper[4967]: I1008 00:17:06.017720 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69dw25g" event={"ID":"277b8c1b-6e5d-44f2-b17b-594c3b367dab","Type":"ContainerDied","Data":"6528f6595e6a3b8c89cedd72e92f2571f2d5bc726915cd1c08392ceedbc137b4"} Oct 08 00:17:06 crc kubenswrapper[4967]: I1008 00:17:06.920471 4967 patch_prober.go:28] interesting pod/machine-config-daemon-xvdtp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 00:17:06 crc kubenswrapper[4967]: I1008 00:17:06.920942 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" podUID="9461162a-c27d-465e-9ca7-2d3936877cd8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 00:17:06 crc kubenswrapper[4967]: I1008 00:17:06.921040 4967 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" Oct 08 00:17:06 crc kubenswrapper[4967]: I1008 00:17:06.921644 4967 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f9c94f180083f2a4c120cfed331389932c804613b3fbe7d131d60b469072cb08"} pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 00:17:06 crc kubenswrapper[4967]: I1008 00:17:06.921770 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" podUID="9461162a-c27d-465e-9ca7-2d3936877cd8" containerName="machine-config-daemon" containerID="cri-o://f9c94f180083f2a4c120cfed331389932c804613b3fbe7d131d60b469072cb08" gracePeriod=600 Oct 08 00:17:06 crc kubenswrapper[4967]: I1008 00:17:06.953037 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/elastic-operator-54896464db-ssjb7"] Oct 08 00:17:06 crc kubenswrapper[4967]: E1008 00:17:06.953234 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="330d296a-13cf-4883-8d5b-af3896f592b8" containerName="pull" Oct 08 00:17:06 crc kubenswrapper[4967]: I1008 00:17:06.953245 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="330d296a-13cf-4883-8d5b-af3896f592b8" containerName="pull" Oct 08 00:17:06 crc kubenswrapper[4967]: E1008 00:17:06.953259 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3221812d-48b9-4361-a75e-3472b715d548" containerName="pull" Oct 08 00:17:06 crc kubenswrapper[4967]: I1008 00:17:06.953266 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="3221812d-48b9-4361-a75e-3472b715d548" containerName="pull" Oct 08 00:17:06 crc kubenswrapper[4967]: E1008 00:17:06.953276 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3221812d-48b9-4361-a75e-3472b715d548" containerName="extract" Oct 08 00:17:06 crc kubenswrapper[4967]: I1008 00:17:06.953282 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="3221812d-48b9-4361-a75e-3472b715d548" containerName="extract" Oct 08 00:17:06 crc kubenswrapper[4967]: E1008 00:17:06.953290 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3221812d-48b9-4361-a75e-3472b715d548" containerName="util" Oct 08 00:17:06 crc kubenswrapper[4967]: I1008 00:17:06.953296 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="3221812d-48b9-4361-a75e-3472b715d548" containerName="util" Oct 08 00:17:06 crc kubenswrapper[4967]: E1008 00:17:06.953305 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="330d296a-13cf-4883-8d5b-af3896f592b8" containerName="util" Oct 08 00:17:06 crc kubenswrapper[4967]: I1008 00:17:06.953310 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="330d296a-13cf-4883-8d5b-af3896f592b8" containerName="util" Oct 08 00:17:06 crc kubenswrapper[4967]: E1008 00:17:06.953320 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="330d296a-13cf-4883-8d5b-af3896f592b8" containerName="extract" Oct 08 00:17:06 crc kubenswrapper[4967]: I1008 00:17:06.953326 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="330d296a-13cf-4883-8d5b-af3896f592b8" containerName="extract" Oct 08 00:17:06 crc kubenswrapper[4967]: I1008 00:17:06.953405 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="3221812d-48b9-4361-a75e-3472b715d548" containerName="extract" Oct 08 00:17:06 crc kubenswrapper[4967]: I1008 00:17:06.953418 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="330d296a-13cf-4883-8d5b-af3896f592b8" containerName="extract" Oct 08 00:17:06 crc kubenswrapper[4967]: I1008 00:17:06.953787 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/elastic-operator-54896464db-ssjb7" Oct 08 00:17:06 crc kubenswrapper[4967]: I1008 00:17:06.955539 4967 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elastic-operator-service-cert" Oct 08 00:17:06 crc kubenswrapper[4967]: I1008 00:17:06.955869 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"kube-root-ca.crt" Oct 08 00:17:06 crc kubenswrapper[4967]: I1008 00:17:06.956884 4967 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elastic-operator-dockercfg-gqzqp" Oct 08 00:17:06 crc kubenswrapper[4967]: I1008 00:17:06.957388 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"openshift-service-ca.crt" Oct 08 00:17:06 crc kubenswrapper[4967]: I1008 00:17:06.967240 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elastic-operator-54896464db-ssjb7"] Oct 08 00:17:07 crc kubenswrapper[4967]: I1008 00:17:07.024385 4967 generic.go:334] "Generic (PLEG): container finished" podID="277b8c1b-6e5d-44f2-b17b-594c3b367dab" containerID="0344d2c0589474bd4eb356c4e1e8a687db55dd2ff30f32e6ab727703ececdcfb" exitCode=0 Oct 08 00:17:07 crc kubenswrapper[4967]: I1008 00:17:07.024426 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69dw25g" event={"ID":"277b8c1b-6e5d-44f2-b17b-594c3b367dab","Type":"ContainerDied","Data":"0344d2c0589474bd4eb356c4e1e8a687db55dd2ff30f32e6ab727703ececdcfb"} Oct 08 00:17:07 crc kubenswrapper[4967]: I1008 00:17:07.093175 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/eb65c076-b2fc-40e4-8d1a-e9084f537d19-webhook-cert\") pod \"elastic-operator-54896464db-ssjb7\" (UID: \"eb65c076-b2fc-40e4-8d1a-e9084f537d19\") " pod="service-telemetry/elastic-operator-54896464db-ssjb7" Oct 08 00:17:07 crc kubenswrapper[4967]: I1008 00:17:07.093224 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-btxbw\" (UniqueName: \"kubernetes.io/projected/eb65c076-b2fc-40e4-8d1a-e9084f537d19-kube-api-access-btxbw\") pod \"elastic-operator-54896464db-ssjb7\" (UID: \"eb65c076-b2fc-40e4-8d1a-e9084f537d19\") " pod="service-telemetry/elastic-operator-54896464db-ssjb7" Oct 08 00:17:07 crc kubenswrapper[4967]: I1008 00:17:07.093279 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/eb65c076-b2fc-40e4-8d1a-e9084f537d19-apiservice-cert\") pod \"elastic-operator-54896464db-ssjb7\" (UID: \"eb65c076-b2fc-40e4-8d1a-e9084f537d19\") " pod="service-telemetry/elastic-operator-54896464db-ssjb7" Oct 08 00:17:07 crc kubenswrapper[4967]: I1008 00:17:07.193964 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/eb65c076-b2fc-40e4-8d1a-e9084f537d19-apiservice-cert\") pod \"elastic-operator-54896464db-ssjb7\" (UID: \"eb65c076-b2fc-40e4-8d1a-e9084f537d19\") " pod="service-telemetry/elastic-operator-54896464db-ssjb7" Oct 08 00:17:07 crc kubenswrapper[4967]: I1008 00:17:07.194063 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/eb65c076-b2fc-40e4-8d1a-e9084f537d19-webhook-cert\") pod \"elastic-operator-54896464db-ssjb7\" (UID: \"eb65c076-b2fc-40e4-8d1a-e9084f537d19\") " pod="service-telemetry/elastic-operator-54896464db-ssjb7" Oct 08 00:17:07 crc kubenswrapper[4967]: I1008 00:17:07.194092 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-btxbw\" (UniqueName: \"kubernetes.io/projected/eb65c076-b2fc-40e4-8d1a-e9084f537d19-kube-api-access-btxbw\") pod \"elastic-operator-54896464db-ssjb7\" (UID: \"eb65c076-b2fc-40e4-8d1a-e9084f537d19\") " pod="service-telemetry/elastic-operator-54896464db-ssjb7" Oct 08 00:17:07 crc kubenswrapper[4967]: I1008 00:17:07.201023 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/eb65c076-b2fc-40e4-8d1a-e9084f537d19-webhook-cert\") pod \"elastic-operator-54896464db-ssjb7\" (UID: \"eb65c076-b2fc-40e4-8d1a-e9084f537d19\") " pod="service-telemetry/elastic-operator-54896464db-ssjb7" Oct 08 00:17:07 crc kubenswrapper[4967]: I1008 00:17:07.204324 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/eb65c076-b2fc-40e4-8d1a-e9084f537d19-apiservice-cert\") pod \"elastic-operator-54896464db-ssjb7\" (UID: \"eb65c076-b2fc-40e4-8d1a-e9084f537d19\") " pod="service-telemetry/elastic-operator-54896464db-ssjb7" Oct 08 00:17:07 crc kubenswrapper[4967]: I1008 00:17:07.221911 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-btxbw\" (UniqueName: \"kubernetes.io/projected/eb65c076-b2fc-40e4-8d1a-e9084f537d19-kube-api-access-btxbw\") pod \"elastic-operator-54896464db-ssjb7\" (UID: \"eb65c076-b2fc-40e4-8d1a-e9084f537d19\") " pod="service-telemetry/elastic-operator-54896464db-ssjb7" Oct 08 00:17:07 crc kubenswrapper[4967]: I1008 00:17:07.309719 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/elastic-operator-54896464db-ssjb7" Oct 08 00:17:07 crc kubenswrapper[4967]: I1008 00:17:07.759711 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elastic-operator-54896464db-ssjb7"] Oct 08 00:17:07 crc kubenswrapper[4967]: W1008 00:17:07.761970 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeb65c076_b2fc_40e4_8d1a_e9084f537d19.slice/crio-3ac0ae30709150ffd4946e7312fba517ba78b117dcb4262d6c290d2582078aed WatchSource:0}: Error finding container 3ac0ae30709150ffd4946e7312fba517ba78b117dcb4262d6c290d2582078aed: Status 404 returned error can't find the container with id 3ac0ae30709150ffd4946e7312fba517ba78b117dcb4262d6c290d2582078aed Oct 08 00:17:08 crc kubenswrapper[4967]: I1008 00:17:08.040990 4967 generic.go:334] "Generic (PLEG): container finished" podID="9461162a-c27d-465e-9ca7-2d3936877cd8" containerID="f9c94f180083f2a4c120cfed331389932c804613b3fbe7d131d60b469072cb08" exitCode=0 Oct 08 00:17:08 crc kubenswrapper[4967]: I1008 00:17:08.041079 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" event={"ID":"9461162a-c27d-465e-9ca7-2d3936877cd8","Type":"ContainerDied","Data":"f9c94f180083f2a4c120cfed331389932c804613b3fbe7d131d60b469072cb08"} Oct 08 00:17:08 crc kubenswrapper[4967]: I1008 00:17:08.041186 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" event={"ID":"9461162a-c27d-465e-9ca7-2d3936877cd8","Type":"ContainerStarted","Data":"312ed5e75849b65841b141a338e52c38e6c54f84a2e110368c62a60516be0c6b"} Oct 08 00:17:08 crc kubenswrapper[4967]: I1008 00:17:08.041248 4967 scope.go:117] "RemoveContainer" containerID="d566016b7da391d83d191f50824299b48d2b961375cdeea4e96f5ac21ff9991f" Oct 08 00:17:08 crc kubenswrapper[4967]: I1008 00:17:08.044177 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elastic-operator-54896464db-ssjb7" event={"ID":"eb65c076-b2fc-40e4-8d1a-e9084f537d19","Type":"ContainerStarted","Data":"3ac0ae30709150ffd4946e7312fba517ba78b117dcb4262d6c290d2582078aed"} Oct 08 00:17:08 crc kubenswrapper[4967]: I1008 00:17:08.320810 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69dw25g" Oct 08 00:17:08 crc kubenswrapper[4967]: I1008 00:17:08.411971 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/277b8c1b-6e5d-44f2-b17b-594c3b367dab-bundle\") pod \"277b8c1b-6e5d-44f2-b17b-594c3b367dab\" (UID: \"277b8c1b-6e5d-44f2-b17b-594c3b367dab\") " Oct 08 00:17:08 crc kubenswrapper[4967]: I1008 00:17:08.412037 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wvnc9\" (UniqueName: \"kubernetes.io/projected/277b8c1b-6e5d-44f2-b17b-594c3b367dab-kube-api-access-wvnc9\") pod \"277b8c1b-6e5d-44f2-b17b-594c3b367dab\" (UID: \"277b8c1b-6e5d-44f2-b17b-594c3b367dab\") " Oct 08 00:17:08 crc kubenswrapper[4967]: I1008 00:17:08.412100 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/277b8c1b-6e5d-44f2-b17b-594c3b367dab-util\") pod \"277b8c1b-6e5d-44f2-b17b-594c3b367dab\" (UID: \"277b8c1b-6e5d-44f2-b17b-594c3b367dab\") " Oct 08 00:17:08 crc kubenswrapper[4967]: I1008 00:17:08.414523 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/277b8c1b-6e5d-44f2-b17b-594c3b367dab-bundle" (OuterVolumeSpecName: "bundle") pod "277b8c1b-6e5d-44f2-b17b-594c3b367dab" (UID: "277b8c1b-6e5d-44f2-b17b-594c3b367dab"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:17:08 crc kubenswrapper[4967]: I1008 00:17:08.424867 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/277b8c1b-6e5d-44f2-b17b-594c3b367dab-kube-api-access-wvnc9" (OuterVolumeSpecName: "kube-api-access-wvnc9") pod "277b8c1b-6e5d-44f2-b17b-594c3b367dab" (UID: "277b8c1b-6e5d-44f2-b17b-594c3b367dab"). InnerVolumeSpecName "kube-api-access-wvnc9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:17:08 crc kubenswrapper[4967]: I1008 00:17:08.432187 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/277b8c1b-6e5d-44f2-b17b-594c3b367dab-util" (OuterVolumeSpecName: "util") pod "277b8c1b-6e5d-44f2-b17b-594c3b367dab" (UID: "277b8c1b-6e5d-44f2-b17b-594c3b367dab"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:17:08 crc kubenswrapper[4967]: I1008 00:17:08.513488 4967 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/277b8c1b-6e5d-44f2-b17b-594c3b367dab-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 00:17:08 crc kubenswrapper[4967]: I1008 00:17:08.513528 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wvnc9\" (UniqueName: \"kubernetes.io/projected/277b8c1b-6e5d-44f2-b17b-594c3b367dab-kube-api-access-wvnc9\") on node \"crc\" DevicePath \"\"" Oct 08 00:17:08 crc kubenswrapper[4967]: I1008 00:17:08.513540 4967 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/277b8c1b-6e5d-44f2-b17b-594c3b367dab-util\") on node \"crc\" DevicePath \"\"" Oct 08 00:17:09 crc kubenswrapper[4967]: I1008 00:17:09.054127 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69dw25g" event={"ID":"277b8c1b-6e5d-44f2-b17b-594c3b367dab","Type":"ContainerDied","Data":"0c880f7673f8747141a82aeb557efba58ac20b3b64b7dbbe98e4db4a52ef8546"} Oct 08 00:17:09 crc kubenswrapper[4967]: I1008 00:17:09.054174 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0c880f7673f8747141a82aeb557efba58ac20b3b64b7dbbe98e4db4a52ef8546" Oct 08 00:17:09 crc kubenswrapper[4967]: I1008 00:17:09.054713 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69dw25g" Oct 08 00:17:11 crc kubenswrapper[4967]: I1008 00:17:11.067101 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elastic-operator-54896464db-ssjb7" event={"ID":"eb65c076-b2fc-40e4-8d1a-e9084f537d19","Type":"ContainerStarted","Data":"9e97cfc3d952f45440124252c49cf5948fc4032527049d31101b4187d2645f75"} Oct 08 00:17:11 crc kubenswrapper[4967]: I1008 00:17:11.088862 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/elastic-operator-54896464db-ssjb7" podStartSLOduration=2.102952933 podStartE2EDuration="5.088842166s" podCreationTimestamp="2025-10-08 00:17:06 +0000 UTC" firstStartedPulling="2025-10-08 00:17:07.774534432 +0000 UTC m=+626.425069091" lastFinishedPulling="2025-10-08 00:17:10.760423675 +0000 UTC m=+629.410958324" observedRunningTime="2025-10-08 00:17:11.086328257 +0000 UTC m=+629.736862926" watchObservedRunningTime="2025-10-08 00:17:11.088842166 +0000 UTC m=+629.739376815" Oct 08 00:17:12 crc kubenswrapper[4967]: I1008 00:17:12.340549 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/interconnect-operator-5bb49f789d-vtczb"] Oct 08 00:17:12 crc kubenswrapper[4967]: E1008 00:17:12.341012 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="277b8c1b-6e5d-44f2-b17b-594c3b367dab" containerName="extract" Oct 08 00:17:12 crc kubenswrapper[4967]: I1008 00:17:12.341024 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="277b8c1b-6e5d-44f2-b17b-594c3b367dab" containerName="extract" Oct 08 00:17:12 crc kubenswrapper[4967]: E1008 00:17:12.341036 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="277b8c1b-6e5d-44f2-b17b-594c3b367dab" containerName="util" Oct 08 00:17:12 crc kubenswrapper[4967]: I1008 00:17:12.341042 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="277b8c1b-6e5d-44f2-b17b-594c3b367dab" containerName="util" Oct 08 00:17:12 crc kubenswrapper[4967]: E1008 00:17:12.341061 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="277b8c1b-6e5d-44f2-b17b-594c3b367dab" containerName="pull" Oct 08 00:17:12 crc kubenswrapper[4967]: I1008 00:17:12.341067 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="277b8c1b-6e5d-44f2-b17b-594c3b367dab" containerName="pull" Oct 08 00:17:12 crc kubenswrapper[4967]: I1008 00:17:12.341152 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="277b8c1b-6e5d-44f2-b17b-594c3b367dab" containerName="extract" Oct 08 00:17:12 crc kubenswrapper[4967]: I1008 00:17:12.341491 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/interconnect-operator-5bb49f789d-vtczb" Oct 08 00:17:12 crc kubenswrapper[4967]: I1008 00:17:12.343303 4967 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"interconnect-operator-dockercfg-jdf8x" Oct 08 00:17:12 crc kubenswrapper[4967]: I1008 00:17:12.356831 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/interconnect-operator-5bb49f789d-vtczb"] Oct 08 00:17:12 crc kubenswrapper[4967]: I1008 00:17:12.470897 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qkpzd\" (UniqueName: \"kubernetes.io/projected/e2154555-42c1-45a2-a30c-70ed7afe09b7-kube-api-access-qkpzd\") pod \"interconnect-operator-5bb49f789d-vtczb\" (UID: \"e2154555-42c1-45a2-a30c-70ed7afe09b7\") " pod="service-telemetry/interconnect-operator-5bb49f789d-vtczb" Oct 08 00:17:12 crc kubenswrapper[4967]: I1008 00:17:12.572339 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qkpzd\" (UniqueName: \"kubernetes.io/projected/e2154555-42c1-45a2-a30c-70ed7afe09b7-kube-api-access-qkpzd\") pod \"interconnect-operator-5bb49f789d-vtczb\" (UID: \"e2154555-42c1-45a2-a30c-70ed7afe09b7\") " pod="service-telemetry/interconnect-operator-5bb49f789d-vtczb" Oct 08 00:17:12 crc kubenswrapper[4967]: I1008 00:17:12.590073 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qkpzd\" (UniqueName: \"kubernetes.io/projected/e2154555-42c1-45a2-a30c-70ed7afe09b7-kube-api-access-qkpzd\") pod \"interconnect-operator-5bb49f789d-vtczb\" (UID: \"e2154555-42c1-45a2-a30c-70ed7afe09b7\") " pod="service-telemetry/interconnect-operator-5bb49f789d-vtczb" Oct 08 00:17:12 crc kubenswrapper[4967]: I1008 00:17:12.655861 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/interconnect-operator-5bb49f789d-vtczb" Oct 08 00:17:13 crc kubenswrapper[4967]: I1008 00:17:13.169188 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/interconnect-operator-5bb49f789d-vtczb"] Oct 08 00:17:13 crc kubenswrapper[4967]: W1008 00:17:13.176910 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode2154555_42c1_45a2_a30c_70ed7afe09b7.slice/crio-33c35f313abb163e335796cefb48818ff0c5dd57bc2de22b278835bdcb8f092a WatchSource:0}: Error finding container 33c35f313abb163e335796cefb48818ff0c5dd57bc2de22b278835bdcb8f092a: Status 404 returned error can't find the container with id 33c35f313abb163e335796cefb48818ff0c5dd57bc2de22b278835bdcb8f092a Oct 08 00:17:14 crc kubenswrapper[4967]: I1008 00:17:14.083519 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/interconnect-operator-5bb49f789d-vtczb" event={"ID":"e2154555-42c1-45a2-a30c-70ed7afe09b7","Type":"ContainerStarted","Data":"33c35f313abb163e335796cefb48818ff0c5dd57bc2de22b278835bdcb8f092a"} Oct 08 00:17:15 crc kubenswrapper[4967]: I1008 00:17:15.621033 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-7c8cf85677-88lmm"] Oct 08 00:17:15 crc kubenswrapper[4967]: I1008 00:17:15.621651 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-88lmm" Oct 08 00:17:15 crc kubenswrapper[4967]: I1008 00:17:15.627190 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Oct 08 00:17:15 crc kubenswrapper[4967]: I1008 00:17:15.628516 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-nr96j" Oct 08 00:17:15 crc kubenswrapper[4967]: I1008 00:17:15.628636 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Oct 08 00:17:15 crc kubenswrapper[4967]: I1008 00:17:15.649777 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-7c8cf85677-88lmm"] Oct 08 00:17:15 crc kubenswrapper[4967]: I1008 00:17:15.708368 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4jql\" (UniqueName: \"kubernetes.io/projected/fb355e35-7b9c-4c0d-bb62-aa169f2bef5a-kube-api-access-z4jql\") pod \"obo-prometheus-operator-7c8cf85677-88lmm\" (UID: \"fb355e35-7b9c-4c0d-bb62-aa169f2bef5a\") " pod="openshift-operators/obo-prometheus-operator-7c8cf85677-88lmm" Oct 08 00:17:15 crc kubenswrapper[4967]: I1008 00:17:15.809955 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4jql\" (UniqueName: \"kubernetes.io/projected/fb355e35-7b9c-4c0d-bb62-aa169f2bef5a-kube-api-access-z4jql\") pod \"obo-prometheus-operator-7c8cf85677-88lmm\" (UID: \"fb355e35-7b9c-4c0d-bb62-aa169f2bef5a\") " pod="openshift-operators/obo-prometheus-operator-7c8cf85677-88lmm" Oct 08 00:17:15 crc kubenswrapper[4967]: I1008 00:17:15.834835 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4jql\" (UniqueName: \"kubernetes.io/projected/fb355e35-7b9c-4c0d-bb62-aa169f2bef5a-kube-api-access-z4jql\") pod \"obo-prometheus-operator-7c8cf85677-88lmm\" (UID: \"fb355e35-7b9c-4c0d-bb62-aa169f2bef5a\") " pod="openshift-operators/obo-prometheus-operator-7c8cf85677-88lmm" Oct 08 00:17:15 crc kubenswrapper[4967]: I1008 00:17:15.900408 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-5fc4456945-q69j8"] Oct 08 00:17:15 crc kubenswrapper[4967]: I1008 00:17:15.901026 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5fc4456945-q69j8" Oct 08 00:17:15 crc kubenswrapper[4967]: I1008 00:17:15.907861 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-fdqs5" Oct 08 00:17:15 crc kubenswrapper[4967]: I1008 00:17:15.914359 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-5fc4456945-kpznr"] Oct 08 00:17:15 crc kubenswrapper[4967]: I1008 00:17:15.915012 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5fc4456945-kpznr" Oct 08 00:17:15 crc kubenswrapper[4967]: I1008 00:17:15.925200 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Oct 08 00:17:15 crc kubenswrapper[4967]: I1008 00:17:15.940299 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-5fc4456945-q69j8"] Oct 08 00:17:15 crc kubenswrapper[4967]: I1008 00:17:15.940691 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-88lmm" Oct 08 00:17:15 crc kubenswrapper[4967]: I1008 00:17:15.963709 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-5fc4456945-kpznr"] Oct 08 00:17:16 crc kubenswrapper[4967]: I1008 00:17:16.013536 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a57260c5-8756-489a-be29-6e71c74778bc-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-5fc4456945-q69j8\" (UID: \"a57260c5-8756-489a-be29-6e71c74778bc\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5fc4456945-q69j8" Oct 08 00:17:16 crc kubenswrapper[4967]: I1008 00:17:16.013612 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2a5fd0cb-9c8e-41e9-af18-d18ddf30180e-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-5fc4456945-kpznr\" (UID: \"2a5fd0cb-9c8e-41e9-af18-d18ddf30180e\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5fc4456945-kpznr" Oct 08 00:17:16 crc kubenswrapper[4967]: I1008 00:17:16.013650 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a57260c5-8756-489a-be29-6e71c74778bc-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-5fc4456945-q69j8\" (UID: \"a57260c5-8756-489a-be29-6e71c74778bc\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5fc4456945-q69j8" Oct 08 00:17:16 crc kubenswrapper[4967]: I1008 00:17:16.013689 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2a5fd0cb-9c8e-41e9-af18-d18ddf30180e-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-5fc4456945-kpznr\" (UID: \"2a5fd0cb-9c8e-41e9-af18-d18ddf30180e\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5fc4456945-kpznr" Oct 08 00:17:16 crc kubenswrapper[4967]: I1008 00:17:16.046409 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-cc5f78dfc-sgk4j"] Oct 08 00:17:16 crc kubenswrapper[4967]: I1008 00:17:16.047182 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-cc5f78dfc-sgk4j" Oct 08 00:17:16 crc kubenswrapper[4967]: I1008 00:17:16.049915 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-j5trp" Oct 08 00:17:16 crc kubenswrapper[4967]: I1008 00:17:16.050236 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Oct 08 00:17:16 crc kubenswrapper[4967]: I1008 00:17:16.064246 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-cc5f78dfc-sgk4j"] Oct 08 00:17:16 crc kubenswrapper[4967]: I1008 00:17:16.114325 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2a5fd0cb-9c8e-41e9-af18-d18ddf30180e-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-5fc4456945-kpznr\" (UID: \"2a5fd0cb-9c8e-41e9-af18-d18ddf30180e\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5fc4456945-kpznr" Oct 08 00:17:16 crc kubenswrapper[4967]: I1008 00:17:16.114390 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z2sv6\" (UniqueName: \"kubernetes.io/projected/d1a621b2-6929-40de-97b1-01a020d70aa2-kube-api-access-z2sv6\") pod \"observability-operator-cc5f78dfc-sgk4j\" (UID: \"d1a621b2-6929-40de-97b1-01a020d70aa2\") " pod="openshift-operators/observability-operator-cc5f78dfc-sgk4j" Oct 08 00:17:16 crc kubenswrapper[4967]: I1008 00:17:16.114415 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a57260c5-8756-489a-be29-6e71c74778bc-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-5fc4456945-q69j8\" (UID: \"a57260c5-8756-489a-be29-6e71c74778bc\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5fc4456945-q69j8" Oct 08 00:17:16 crc kubenswrapper[4967]: I1008 00:17:16.114454 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2a5fd0cb-9c8e-41e9-af18-d18ddf30180e-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-5fc4456945-kpznr\" (UID: \"2a5fd0cb-9c8e-41e9-af18-d18ddf30180e\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5fc4456945-kpznr" Oct 08 00:17:16 crc kubenswrapper[4967]: I1008 00:17:16.114473 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/d1a621b2-6929-40de-97b1-01a020d70aa2-observability-operator-tls\") pod \"observability-operator-cc5f78dfc-sgk4j\" (UID: \"d1a621b2-6929-40de-97b1-01a020d70aa2\") " pod="openshift-operators/observability-operator-cc5f78dfc-sgk4j" Oct 08 00:17:16 crc kubenswrapper[4967]: I1008 00:17:16.114499 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a57260c5-8756-489a-be29-6e71c74778bc-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-5fc4456945-q69j8\" (UID: \"a57260c5-8756-489a-be29-6e71c74778bc\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5fc4456945-q69j8" Oct 08 00:17:16 crc kubenswrapper[4967]: I1008 00:17:16.118717 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a57260c5-8756-489a-be29-6e71c74778bc-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-5fc4456945-q69j8\" (UID: \"a57260c5-8756-489a-be29-6e71c74778bc\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5fc4456945-q69j8" Oct 08 00:17:16 crc kubenswrapper[4967]: I1008 00:17:16.119493 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2a5fd0cb-9c8e-41e9-af18-d18ddf30180e-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-5fc4456945-kpznr\" (UID: \"2a5fd0cb-9c8e-41e9-af18-d18ddf30180e\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5fc4456945-kpznr" Oct 08 00:17:16 crc kubenswrapper[4967]: I1008 00:17:16.121978 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a57260c5-8756-489a-be29-6e71c74778bc-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-5fc4456945-q69j8\" (UID: \"a57260c5-8756-489a-be29-6e71c74778bc\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5fc4456945-q69j8" Oct 08 00:17:16 crc kubenswrapper[4967]: I1008 00:17:16.133131 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2a5fd0cb-9c8e-41e9-af18-d18ddf30180e-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-5fc4456945-kpznr\" (UID: \"2a5fd0cb-9c8e-41e9-af18-d18ddf30180e\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5fc4456945-kpznr" Oct 08 00:17:16 crc kubenswrapper[4967]: I1008 00:17:16.217267 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z2sv6\" (UniqueName: \"kubernetes.io/projected/d1a621b2-6929-40de-97b1-01a020d70aa2-kube-api-access-z2sv6\") pod \"observability-operator-cc5f78dfc-sgk4j\" (UID: \"d1a621b2-6929-40de-97b1-01a020d70aa2\") " pod="openshift-operators/observability-operator-cc5f78dfc-sgk4j" Oct 08 00:17:16 crc kubenswrapper[4967]: I1008 00:17:16.217581 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/d1a621b2-6929-40de-97b1-01a020d70aa2-observability-operator-tls\") pod \"observability-operator-cc5f78dfc-sgk4j\" (UID: \"d1a621b2-6929-40de-97b1-01a020d70aa2\") " pod="openshift-operators/observability-operator-cc5f78dfc-sgk4j" Oct 08 00:17:16 crc kubenswrapper[4967]: I1008 00:17:16.225061 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5fc4456945-q69j8" Oct 08 00:17:16 crc kubenswrapper[4967]: I1008 00:17:16.230388 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/d1a621b2-6929-40de-97b1-01a020d70aa2-observability-operator-tls\") pod \"observability-operator-cc5f78dfc-sgk4j\" (UID: \"d1a621b2-6929-40de-97b1-01a020d70aa2\") " pod="openshift-operators/observability-operator-cc5f78dfc-sgk4j" Oct 08 00:17:16 crc kubenswrapper[4967]: I1008 00:17:16.248996 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5fc4456945-kpznr" Oct 08 00:17:16 crc kubenswrapper[4967]: I1008 00:17:16.254724 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z2sv6\" (UniqueName: \"kubernetes.io/projected/d1a621b2-6929-40de-97b1-01a020d70aa2-kube-api-access-z2sv6\") pod \"observability-operator-cc5f78dfc-sgk4j\" (UID: \"d1a621b2-6929-40de-97b1-01a020d70aa2\") " pod="openshift-operators/observability-operator-cc5f78dfc-sgk4j" Oct 08 00:17:16 crc kubenswrapper[4967]: I1008 00:17:16.258378 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-54bc95c9fb-x476n"] Oct 08 00:17:16 crc kubenswrapper[4967]: I1008 00:17:16.270105 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-54bc95c9fb-x476n" Oct 08 00:17:16 crc kubenswrapper[4967]: I1008 00:17:16.286869 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-54bc95c9fb-x476n"] Oct 08 00:17:16 crc kubenswrapper[4967]: I1008 00:17:16.287585 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-g6htg" Oct 08 00:17:16 crc kubenswrapper[4967]: I1008 00:17:16.322241 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6vwgm\" (UniqueName: \"kubernetes.io/projected/478b5edf-299e-4d19-bc8f-02a5e9774148-kube-api-access-6vwgm\") pod \"perses-operator-54bc95c9fb-x476n\" (UID: \"478b5edf-299e-4d19-bc8f-02a5e9774148\") " pod="openshift-operators/perses-operator-54bc95c9fb-x476n" Oct 08 00:17:16 crc kubenswrapper[4967]: I1008 00:17:16.322312 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/478b5edf-299e-4d19-bc8f-02a5e9774148-openshift-service-ca\") pod \"perses-operator-54bc95c9fb-x476n\" (UID: \"478b5edf-299e-4d19-bc8f-02a5e9774148\") " pod="openshift-operators/perses-operator-54bc95c9fb-x476n" Oct 08 00:17:16 crc kubenswrapper[4967]: I1008 00:17:16.378051 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-cc5f78dfc-sgk4j" Oct 08 00:17:16 crc kubenswrapper[4967]: I1008 00:17:16.426245 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6vwgm\" (UniqueName: \"kubernetes.io/projected/478b5edf-299e-4d19-bc8f-02a5e9774148-kube-api-access-6vwgm\") pod \"perses-operator-54bc95c9fb-x476n\" (UID: \"478b5edf-299e-4d19-bc8f-02a5e9774148\") " pod="openshift-operators/perses-operator-54bc95c9fb-x476n" Oct 08 00:17:16 crc kubenswrapper[4967]: I1008 00:17:16.426307 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/478b5edf-299e-4d19-bc8f-02a5e9774148-openshift-service-ca\") pod \"perses-operator-54bc95c9fb-x476n\" (UID: \"478b5edf-299e-4d19-bc8f-02a5e9774148\") " pod="openshift-operators/perses-operator-54bc95c9fb-x476n" Oct 08 00:17:16 crc kubenswrapper[4967]: I1008 00:17:16.427333 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/478b5edf-299e-4d19-bc8f-02a5e9774148-openshift-service-ca\") pod \"perses-operator-54bc95c9fb-x476n\" (UID: \"478b5edf-299e-4d19-bc8f-02a5e9774148\") " pod="openshift-operators/perses-operator-54bc95c9fb-x476n" Oct 08 00:17:16 crc kubenswrapper[4967]: I1008 00:17:16.452336 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6vwgm\" (UniqueName: \"kubernetes.io/projected/478b5edf-299e-4d19-bc8f-02a5e9774148-kube-api-access-6vwgm\") pod \"perses-operator-54bc95c9fb-x476n\" (UID: \"478b5edf-299e-4d19-bc8f-02a5e9774148\") " pod="openshift-operators/perses-operator-54bc95c9fb-x476n" Oct 08 00:17:16 crc kubenswrapper[4967]: I1008 00:17:16.480091 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-7c8cf85677-88lmm"] Oct 08 00:17:16 crc kubenswrapper[4967]: W1008 00:17:16.515087 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfb355e35_7b9c_4c0d_bb62_aa169f2bef5a.slice/crio-79debec6db059a9717a1600c008f0e701e4440caf44a55d754dd772e568640ea WatchSource:0}: Error finding container 79debec6db059a9717a1600c008f0e701e4440caf44a55d754dd772e568640ea: Status 404 returned error can't find the container with id 79debec6db059a9717a1600c008f0e701e4440caf44a55d754dd772e568640ea Oct 08 00:17:16 crc kubenswrapper[4967]: I1008 00:17:16.660102 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-54bc95c9fb-x476n" Oct 08 00:17:16 crc kubenswrapper[4967]: I1008 00:17:16.829280 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-5fc4456945-q69j8"] Oct 08 00:17:16 crc kubenswrapper[4967]: I1008 00:17:16.965175 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-cc5f78dfc-sgk4j"] Oct 08 00:17:16 crc kubenswrapper[4967]: W1008 00:17:16.975818 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1a621b2_6929_40de_97b1_01a020d70aa2.slice/crio-e96e4ce29904ba385c6cd47bc8a848bf84be7122bbd6fa2e3ccc813a43c5d4fb WatchSource:0}: Error finding container e96e4ce29904ba385c6cd47bc8a848bf84be7122bbd6fa2e3ccc813a43c5d4fb: Status 404 returned error can't find the container with id e96e4ce29904ba385c6cd47bc8a848bf84be7122bbd6fa2e3ccc813a43c5d4fb Oct 08 00:17:16 crc kubenswrapper[4967]: I1008 00:17:16.976839 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-5fc4456945-kpznr"] Oct 08 00:17:16 crc kubenswrapper[4967]: W1008 00:17:16.985448 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2a5fd0cb_9c8e_41e9_af18_d18ddf30180e.slice/crio-1f288acade94ced7c6c7fbe0cfe318af5b730dc45c6317d4ebeb88932bf0ba67 WatchSource:0}: Error finding container 1f288acade94ced7c6c7fbe0cfe318af5b730dc45c6317d4ebeb88932bf0ba67: Status 404 returned error can't find the container with id 1f288acade94ced7c6c7fbe0cfe318af5b730dc45c6317d4ebeb88932bf0ba67 Oct 08 00:17:17 crc kubenswrapper[4967]: I1008 00:17:17.082932 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-54bc95c9fb-x476n"] Oct 08 00:17:17 crc kubenswrapper[4967]: I1008 00:17:17.152741 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5fc4456945-q69j8" event={"ID":"a57260c5-8756-489a-be29-6e71c74778bc","Type":"ContainerStarted","Data":"e8b13ec8d74a23d9a5d52d8243b5fd3754d86c64211c1f2deff33b4fb65e59d1"} Oct 08 00:17:17 crc kubenswrapper[4967]: I1008 00:17:17.155433 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5fc4456945-kpznr" event={"ID":"2a5fd0cb-9c8e-41e9-af18-d18ddf30180e","Type":"ContainerStarted","Data":"1f288acade94ced7c6c7fbe0cfe318af5b730dc45c6317d4ebeb88932bf0ba67"} Oct 08 00:17:17 crc kubenswrapper[4967]: I1008 00:17:17.158704 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-54bc95c9fb-x476n" event={"ID":"478b5edf-299e-4d19-bc8f-02a5e9774148","Type":"ContainerStarted","Data":"ac70e4e2fed764a1663f0ca8b66f26b5b63db7eb47f1b2398f557dd028109da7"} Oct 08 00:17:17 crc kubenswrapper[4967]: I1008 00:17:17.159777 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-cc5f78dfc-sgk4j" event={"ID":"d1a621b2-6929-40de-97b1-01a020d70aa2","Type":"ContainerStarted","Data":"e96e4ce29904ba385c6cd47bc8a848bf84be7122bbd6fa2e3ccc813a43c5d4fb"} Oct 08 00:17:17 crc kubenswrapper[4967]: I1008 00:17:17.168502 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-88lmm" event={"ID":"fb355e35-7b9c-4c0d-bb62-aa169f2bef5a","Type":"ContainerStarted","Data":"79debec6db059a9717a1600c008f0e701e4440caf44a55d754dd772e568640ea"} Oct 08 00:17:24 crc kubenswrapper[4967]: I1008 00:17:24.607383 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-96d66748b-w49xk"] Oct 08 00:17:24 crc kubenswrapper[4967]: I1008 00:17:24.608362 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-96d66748b-w49xk" Oct 08 00:17:24 crc kubenswrapper[4967]: I1008 00:17:24.610772 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"openshift-service-ca.crt" Oct 08 00:17:24 crc kubenswrapper[4967]: I1008 00:17:24.611290 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"kube-root-ca.crt" Oct 08 00:17:24 crc kubenswrapper[4967]: I1008 00:17:24.611538 4967 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager-operator"/"cert-manager-operator-controller-manager-dockercfg-b2bpt" Oct 08 00:17:24 crc kubenswrapper[4967]: I1008 00:17:24.626994 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-96d66748b-w49xk"] Oct 08 00:17:24 crc kubenswrapper[4967]: I1008 00:17:24.656742 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fdtrc\" (UniqueName: \"kubernetes.io/projected/83068de3-9ba5-4672-b843-b36adba4916e-kube-api-access-fdtrc\") pod \"cert-manager-operator-controller-manager-96d66748b-w49xk\" (UID: \"83068de3-9ba5-4672-b843-b36adba4916e\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-96d66748b-w49xk" Oct 08 00:17:24 crc kubenswrapper[4967]: I1008 00:17:24.757821 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fdtrc\" (UniqueName: \"kubernetes.io/projected/83068de3-9ba5-4672-b843-b36adba4916e-kube-api-access-fdtrc\") pod \"cert-manager-operator-controller-manager-96d66748b-w49xk\" (UID: \"83068de3-9ba5-4672-b843-b36adba4916e\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-96d66748b-w49xk" Oct 08 00:17:24 crc kubenswrapper[4967]: I1008 00:17:24.777468 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fdtrc\" (UniqueName: \"kubernetes.io/projected/83068de3-9ba5-4672-b843-b36adba4916e-kube-api-access-fdtrc\") pod \"cert-manager-operator-controller-manager-96d66748b-w49xk\" (UID: \"83068de3-9ba5-4672-b843-b36adba4916e\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-96d66748b-w49xk" Oct 08 00:17:24 crc kubenswrapper[4967]: I1008 00:17:24.922177 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-96d66748b-w49xk" Oct 08 00:17:25 crc kubenswrapper[4967]: I1008 00:17:25.169662 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Oct 08 00:17:25 crc kubenswrapper[4967]: I1008 00:17:25.170691 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/elasticsearch-es-default-0" Oct 08 00:17:25 crc kubenswrapper[4967]: I1008 00:17:25.176065 4967 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-internal-users" Oct 08 00:17:25 crc kubenswrapper[4967]: I1008 00:17:25.176062 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"elasticsearch-es-unicast-hosts" Oct 08 00:17:25 crc kubenswrapper[4967]: I1008 00:17:25.176165 4967 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-remote-ca" Oct 08 00:17:25 crc kubenswrapper[4967]: I1008 00:17:25.176401 4967 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-dockercfg-jnqfs" Oct 08 00:17:25 crc kubenswrapper[4967]: I1008 00:17:25.176658 4967 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-default-es-transport-certs" Oct 08 00:17:25 crc kubenswrapper[4967]: I1008 00:17:25.176928 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"elasticsearch-es-scripts" Oct 08 00:17:25 crc kubenswrapper[4967]: I1008 00:17:25.177106 4967 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-default-es-config" Oct 08 00:17:25 crc kubenswrapper[4967]: I1008 00:17:25.177946 4967 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-http-certs-internal" Oct 08 00:17:25 crc kubenswrapper[4967]: I1008 00:17:25.181516 4967 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-xpack-file-realm" Oct 08 00:17:25 crc kubenswrapper[4967]: I1008 00:17:25.198273 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Oct 08 00:17:25 crc kubenswrapper[4967]: I1008 00:17:25.263710 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-elasticsearch-config\" (UniqueName: \"kubernetes.io/secret/79b69a1c-270d-4261-b80f-3aff7be1c7bf-elastic-internal-elasticsearch-config\") pod \"elasticsearch-es-default-0\" (UID: \"79b69a1c-270d-4261-b80f-3aff7be1c7bf\") " pod="service-telemetry/elasticsearch-es-default-0" Oct 08 00:17:25 crc kubenswrapper[4967]: I1008 00:17:25.263777 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-http-certificates\" (UniqueName: \"kubernetes.io/secret/79b69a1c-270d-4261-b80f-3aff7be1c7bf-elastic-internal-http-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"79b69a1c-270d-4261-b80f-3aff7be1c7bf\") " pod="service-telemetry/elasticsearch-es-default-0" Oct 08 00:17:25 crc kubenswrapper[4967]: I1008 00:17:25.263819 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-elasticsearch-config-local\" (UniqueName: \"kubernetes.io/empty-dir/79b69a1c-270d-4261-b80f-3aff7be1c7bf-elastic-internal-elasticsearch-config-local\") pod \"elasticsearch-es-default-0\" (UID: \"79b69a1c-270d-4261-b80f-3aff7be1c7bf\") " pod="service-telemetry/elasticsearch-es-default-0" Oct 08 00:17:25 crc kubenswrapper[4967]: I1008 00:17:25.263839 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elasticsearch-logs\" (UniqueName: \"kubernetes.io/empty-dir/79b69a1c-270d-4261-b80f-3aff7be1c7bf-elasticsearch-logs\") pod \"elasticsearch-es-default-0\" (UID: \"79b69a1c-270d-4261-b80f-3aff7be1c7bf\") " pod="service-telemetry/elasticsearch-es-default-0" Oct 08 00:17:25 crc kubenswrapper[4967]: I1008 00:17:25.263984 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp-volume\" (UniqueName: \"kubernetes.io/empty-dir/79b69a1c-270d-4261-b80f-3aff7be1c7bf-tmp-volume\") pod \"elasticsearch-es-default-0\" (UID: \"79b69a1c-270d-4261-b80f-3aff7be1c7bf\") " pod="service-telemetry/elasticsearch-es-default-0" Oct 08 00:17:25 crc kubenswrapper[4967]: I1008 00:17:25.264040 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-elasticsearch-bin-local\" (UniqueName: \"kubernetes.io/empty-dir/79b69a1c-270d-4261-b80f-3aff7be1c7bf-elastic-internal-elasticsearch-bin-local\") pod \"elasticsearch-es-default-0\" (UID: \"79b69a1c-270d-4261-b80f-3aff7be1c7bf\") " pod="service-telemetry/elasticsearch-es-default-0" Oct 08 00:17:25 crc kubenswrapper[4967]: I1008 00:17:25.264112 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-unicast-hosts\" (UniqueName: \"kubernetes.io/configmap/79b69a1c-270d-4261-b80f-3aff7be1c7bf-elastic-internal-unicast-hosts\") pod \"elasticsearch-es-default-0\" (UID: \"79b69a1c-270d-4261-b80f-3aff7be1c7bf\") " pod="service-telemetry/elasticsearch-es-default-0" Oct 08 00:17:25 crc kubenswrapper[4967]: I1008 00:17:25.264193 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-elasticsearch-plugins-local\" (UniqueName: \"kubernetes.io/empty-dir/79b69a1c-270d-4261-b80f-3aff7be1c7bf-elastic-internal-elasticsearch-plugins-local\") pod \"elasticsearch-es-default-0\" (UID: \"79b69a1c-270d-4261-b80f-3aff7be1c7bf\") " pod="service-telemetry/elasticsearch-es-default-0" Oct 08 00:17:25 crc kubenswrapper[4967]: I1008 00:17:25.264221 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elasticsearch-data\" (UniqueName: \"kubernetes.io/empty-dir/79b69a1c-270d-4261-b80f-3aff7be1c7bf-elasticsearch-data\") pod \"elasticsearch-es-default-0\" (UID: \"79b69a1c-270d-4261-b80f-3aff7be1c7bf\") " pod="service-telemetry/elasticsearch-es-default-0" Oct 08 00:17:25 crc kubenswrapper[4967]: I1008 00:17:25.264340 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-probe-user\" (UniqueName: \"kubernetes.io/secret/79b69a1c-270d-4261-b80f-3aff7be1c7bf-elastic-internal-probe-user\") pod \"elasticsearch-es-default-0\" (UID: \"79b69a1c-270d-4261-b80f-3aff7be1c7bf\") " pod="service-telemetry/elasticsearch-es-default-0" Oct 08 00:17:25 crc kubenswrapper[4967]: I1008 00:17:25.264406 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-scripts\" (UniqueName: \"kubernetes.io/configmap/79b69a1c-270d-4261-b80f-3aff7be1c7bf-elastic-internal-scripts\") pod \"elasticsearch-es-default-0\" (UID: \"79b69a1c-270d-4261-b80f-3aff7be1c7bf\") " pod="service-telemetry/elasticsearch-es-default-0" Oct 08 00:17:25 crc kubenswrapper[4967]: I1008 00:17:25.264448 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-remote-certificate-authorities\" (UniqueName: \"kubernetes.io/secret/79b69a1c-270d-4261-b80f-3aff7be1c7bf-elastic-internal-remote-certificate-authorities\") pod \"elasticsearch-es-default-0\" (UID: \"79b69a1c-270d-4261-b80f-3aff7be1c7bf\") " pod="service-telemetry/elasticsearch-es-default-0" Oct 08 00:17:25 crc kubenswrapper[4967]: I1008 00:17:25.264480 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"downward-api\" (UniqueName: \"kubernetes.io/downward-api/79b69a1c-270d-4261-b80f-3aff7be1c7bf-downward-api\") pod \"elasticsearch-es-default-0\" (UID: \"79b69a1c-270d-4261-b80f-3aff7be1c7bf\") " pod="service-telemetry/elasticsearch-es-default-0" Oct 08 00:17:25 crc kubenswrapper[4967]: I1008 00:17:25.264516 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-xpack-file-realm\" (UniqueName: \"kubernetes.io/secret/79b69a1c-270d-4261-b80f-3aff7be1c7bf-elastic-internal-xpack-file-realm\") pod \"elasticsearch-es-default-0\" (UID: \"79b69a1c-270d-4261-b80f-3aff7be1c7bf\") " pod="service-telemetry/elasticsearch-es-default-0" Oct 08 00:17:25 crc kubenswrapper[4967]: I1008 00:17:25.264543 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-transport-certificates\" (UniqueName: \"kubernetes.io/secret/79b69a1c-270d-4261-b80f-3aff7be1c7bf-elastic-internal-transport-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"79b69a1c-270d-4261-b80f-3aff7be1c7bf\") " pod="service-telemetry/elasticsearch-es-default-0" Oct 08 00:17:25 crc kubenswrapper[4967]: I1008 00:17:25.365821 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-elasticsearch-plugins-local\" (UniqueName: \"kubernetes.io/empty-dir/79b69a1c-270d-4261-b80f-3aff7be1c7bf-elastic-internal-elasticsearch-plugins-local\") pod \"elasticsearch-es-default-0\" (UID: \"79b69a1c-270d-4261-b80f-3aff7be1c7bf\") " pod="service-telemetry/elasticsearch-es-default-0" Oct 08 00:17:25 crc kubenswrapper[4967]: I1008 00:17:25.365867 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elasticsearch-data\" (UniqueName: \"kubernetes.io/empty-dir/79b69a1c-270d-4261-b80f-3aff7be1c7bf-elasticsearch-data\") pod \"elasticsearch-es-default-0\" (UID: \"79b69a1c-270d-4261-b80f-3aff7be1c7bf\") " pod="service-telemetry/elasticsearch-es-default-0" Oct 08 00:17:25 crc kubenswrapper[4967]: I1008 00:17:25.365886 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-probe-user\" (UniqueName: \"kubernetes.io/secret/79b69a1c-270d-4261-b80f-3aff7be1c7bf-elastic-internal-probe-user\") pod \"elasticsearch-es-default-0\" (UID: \"79b69a1c-270d-4261-b80f-3aff7be1c7bf\") " pod="service-telemetry/elasticsearch-es-default-0" Oct 08 00:17:25 crc kubenswrapper[4967]: I1008 00:17:25.365908 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-scripts\" (UniqueName: \"kubernetes.io/configmap/79b69a1c-270d-4261-b80f-3aff7be1c7bf-elastic-internal-scripts\") pod \"elasticsearch-es-default-0\" (UID: \"79b69a1c-270d-4261-b80f-3aff7be1c7bf\") " pod="service-telemetry/elasticsearch-es-default-0" Oct 08 00:17:25 crc kubenswrapper[4967]: I1008 00:17:25.365929 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-remote-certificate-authorities\" (UniqueName: \"kubernetes.io/secret/79b69a1c-270d-4261-b80f-3aff7be1c7bf-elastic-internal-remote-certificate-authorities\") pod \"elasticsearch-es-default-0\" (UID: \"79b69a1c-270d-4261-b80f-3aff7be1c7bf\") " pod="service-telemetry/elasticsearch-es-default-0" Oct 08 00:17:25 crc kubenswrapper[4967]: I1008 00:17:25.365970 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"downward-api\" (UniqueName: \"kubernetes.io/downward-api/79b69a1c-270d-4261-b80f-3aff7be1c7bf-downward-api\") pod \"elasticsearch-es-default-0\" (UID: \"79b69a1c-270d-4261-b80f-3aff7be1c7bf\") " pod="service-telemetry/elasticsearch-es-default-0" Oct 08 00:17:25 crc kubenswrapper[4967]: I1008 00:17:25.365986 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-xpack-file-realm\" (UniqueName: \"kubernetes.io/secret/79b69a1c-270d-4261-b80f-3aff7be1c7bf-elastic-internal-xpack-file-realm\") pod \"elasticsearch-es-default-0\" (UID: \"79b69a1c-270d-4261-b80f-3aff7be1c7bf\") " pod="service-telemetry/elasticsearch-es-default-0" Oct 08 00:17:25 crc kubenswrapper[4967]: I1008 00:17:25.366006 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-transport-certificates\" (UniqueName: \"kubernetes.io/secret/79b69a1c-270d-4261-b80f-3aff7be1c7bf-elastic-internal-transport-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"79b69a1c-270d-4261-b80f-3aff7be1c7bf\") " pod="service-telemetry/elasticsearch-es-default-0" Oct 08 00:17:25 crc kubenswrapper[4967]: I1008 00:17:25.366041 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-elasticsearch-config\" (UniqueName: \"kubernetes.io/secret/79b69a1c-270d-4261-b80f-3aff7be1c7bf-elastic-internal-elasticsearch-config\") pod \"elasticsearch-es-default-0\" (UID: \"79b69a1c-270d-4261-b80f-3aff7be1c7bf\") " pod="service-telemetry/elasticsearch-es-default-0" Oct 08 00:17:25 crc kubenswrapper[4967]: I1008 00:17:25.366060 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-http-certificates\" (UniqueName: \"kubernetes.io/secret/79b69a1c-270d-4261-b80f-3aff7be1c7bf-elastic-internal-http-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"79b69a1c-270d-4261-b80f-3aff7be1c7bf\") " pod="service-telemetry/elasticsearch-es-default-0" Oct 08 00:17:25 crc kubenswrapper[4967]: I1008 00:17:25.366086 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elasticsearch-logs\" (UniqueName: \"kubernetes.io/empty-dir/79b69a1c-270d-4261-b80f-3aff7be1c7bf-elasticsearch-logs\") pod \"elasticsearch-es-default-0\" (UID: \"79b69a1c-270d-4261-b80f-3aff7be1c7bf\") " pod="service-telemetry/elasticsearch-es-default-0" Oct 08 00:17:25 crc kubenswrapper[4967]: I1008 00:17:25.366103 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-elasticsearch-config-local\" (UniqueName: \"kubernetes.io/empty-dir/79b69a1c-270d-4261-b80f-3aff7be1c7bf-elastic-internal-elasticsearch-config-local\") pod \"elasticsearch-es-default-0\" (UID: \"79b69a1c-270d-4261-b80f-3aff7be1c7bf\") " pod="service-telemetry/elasticsearch-es-default-0" Oct 08 00:17:25 crc kubenswrapper[4967]: I1008 00:17:25.366119 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp-volume\" (UniqueName: \"kubernetes.io/empty-dir/79b69a1c-270d-4261-b80f-3aff7be1c7bf-tmp-volume\") pod \"elasticsearch-es-default-0\" (UID: \"79b69a1c-270d-4261-b80f-3aff7be1c7bf\") " pod="service-telemetry/elasticsearch-es-default-0" Oct 08 00:17:25 crc kubenswrapper[4967]: I1008 00:17:25.366135 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-elasticsearch-bin-local\" (UniqueName: \"kubernetes.io/empty-dir/79b69a1c-270d-4261-b80f-3aff7be1c7bf-elastic-internal-elasticsearch-bin-local\") pod \"elasticsearch-es-default-0\" (UID: \"79b69a1c-270d-4261-b80f-3aff7be1c7bf\") " pod="service-telemetry/elasticsearch-es-default-0" Oct 08 00:17:25 crc kubenswrapper[4967]: I1008 00:17:25.366159 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-unicast-hosts\" (UniqueName: \"kubernetes.io/configmap/79b69a1c-270d-4261-b80f-3aff7be1c7bf-elastic-internal-unicast-hosts\") pod \"elasticsearch-es-default-0\" (UID: \"79b69a1c-270d-4261-b80f-3aff7be1c7bf\") " pod="service-telemetry/elasticsearch-es-default-0" Oct 08 00:17:25 crc kubenswrapper[4967]: I1008 00:17:25.366799 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-elasticsearch-plugins-local\" (UniqueName: \"kubernetes.io/empty-dir/79b69a1c-270d-4261-b80f-3aff7be1c7bf-elastic-internal-elasticsearch-plugins-local\") pod \"elasticsearch-es-default-0\" (UID: \"79b69a1c-270d-4261-b80f-3aff7be1c7bf\") " pod="service-telemetry/elasticsearch-es-default-0" Oct 08 00:17:25 crc kubenswrapper[4967]: I1008 00:17:25.369880 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-elasticsearch-config-local\" (UniqueName: \"kubernetes.io/empty-dir/79b69a1c-270d-4261-b80f-3aff7be1c7bf-elastic-internal-elasticsearch-config-local\") pod \"elasticsearch-es-default-0\" (UID: \"79b69a1c-270d-4261-b80f-3aff7be1c7bf\") " pod="service-telemetry/elasticsearch-es-default-0" Oct 08 00:17:25 crc kubenswrapper[4967]: I1008 00:17:25.380033 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elasticsearch-logs\" (UniqueName: \"kubernetes.io/empty-dir/79b69a1c-270d-4261-b80f-3aff7be1c7bf-elasticsearch-logs\") pod \"elasticsearch-es-default-0\" (UID: \"79b69a1c-270d-4261-b80f-3aff7be1c7bf\") " pod="service-telemetry/elasticsearch-es-default-0" Oct 08 00:17:25 crc kubenswrapper[4967]: I1008 00:17:25.380043 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp-volume\" (UniqueName: \"kubernetes.io/empty-dir/79b69a1c-270d-4261-b80f-3aff7be1c7bf-tmp-volume\") pod \"elasticsearch-es-default-0\" (UID: \"79b69a1c-270d-4261-b80f-3aff7be1c7bf\") " pod="service-telemetry/elasticsearch-es-default-0" Oct 08 00:17:25 crc kubenswrapper[4967]: I1008 00:17:25.382298 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-unicast-hosts\" (UniqueName: \"kubernetes.io/configmap/79b69a1c-270d-4261-b80f-3aff7be1c7bf-elastic-internal-unicast-hosts\") pod \"elasticsearch-es-default-0\" (UID: \"79b69a1c-270d-4261-b80f-3aff7be1c7bf\") " pod="service-telemetry/elasticsearch-es-default-0" Oct 08 00:17:25 crc kubenswrapper[4967]: I1008 00:17:25.388359 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-probe-user\" (UniqueName: \"kubernetes.io/secret/79b69a1c-270d-4261-b80f-3aff7be1c7bf-elastic-internal-probe-user\") pod \"elasticsearch-es-default-0\" (UID: \"79b69a1c-270d-4261-b80f-3aff7be1c7bf\") " pod="service-telemetry/elasticsearch-es-default-0" Oct 08 00:17:25 crc kubenswrapper[4967]: I1008 00:17:25.389746 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-xpack-file-realm\" (UniqueName: \"kubernetes.io/secret/79b69a1c-270d-4261-b80f-3aff7be1c7bf-elastic-internal-xpack-file-realm\") pod \"elasticsearch-es-default-0\" (UID: \"79b69a1c-270d-4261-b80f-3aff7be1c7bf\") " pod="service-telemetry/elasticsearch-es-default-0" Oct 08 00:17:25 crc kubenswrapper[4967]: I1008 00:17:25.389898 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elasticsearch-data\" (UniqueName: \"kubernetes.io/empty-dir/79b69a1c-270d-4261-b80f-3aff7be1c7bf-elasticsearch-data\") pod \"elasticsearch-es-default-0\" (UID: \"79b69a1c-270d-4261-b80f-3aff7be1c7bf\") " pod="service-telemetry/elasticsearch-es-default-0" Oct 08 00:17:25 crc kubenswrapper[4967]: I1008 00:17:25.391346 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-transport-certificates\" (UniqueName: \"kubernetes.io/secret/79b69a1c-270d-4261-b80f-3aff7be1c7bf-elastic-internal-transport-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"79b69a1c-270d-4261-b80f-3aff7be1c7bf\") " pod="service-telemetry/elasticsearch-es-default-0" Oct 08 00:17:25 crc kubenswrapper[4967]: I1008 00:17:25.393478 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-scripts\" (UniqueName: \"kubernetes.io/configmap/79b69a1c-270d-4261-b80f-3aff7be1c7bf-elastic-internal-scripts\") pod \"elasticsearch-es-default-0\" (UID: \"79b69a1c-270d-4261-b80f-3aff7be1c7bf\") " pod="service-telemetry/elasticsearch-es-default-0" Oct 08 00:17:25 crc kubenswrapper[4967]: I1008 00:17:25.393892 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-elasticsearch-config\" (UniqueName: \"kubernetes.io/secret/79b69a1c-270d-4261-b80f-3aff7be1c7bf-elastic-internal-elasticsearch-config\") pod \"elasticsearch-es-default-0\" (UID: \"79b69a1c-270d-4261-b80f-3aff7be1c7bf\") " pod="service-telemetry/elasticsearch-es-default-0" Oct 08 00:17:25 crc kubenswrapper[4967]: I1008 00:17:25.399167 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-http-certificates\" (UniqueName: \"kubernetes.io/secret/79b69a1c-270d-4261-b80f-3aff7be1c7bf-elastic-internal-http-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"79b69a1c-270d-4261-b80f-3aff7be1c7bf\") " pod="service-telemetry/elasticsearch-es-default-0" Oct 08 00:17:25 crc kubenswrapper[4967]: I1008 00:17:25.400395 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-remote-certificate-authorities\" (UniqueName: \"kubernetes.io/secret/79b69a1c-270d-4261-b80f-3aff7be1c7bf-elastic-internal-remote-certificate-authorities\") pod \"elasticsearch-es-default-0\" (UID: \"79b69a1c-270d-4261-b80f-3aff7be1c7bf\") " pod="service-telemetry/elasticsearch-es-default-0" Oct 08 00:17:25 crc kubenswrapper[4967]: I1008 00:17:25.402232 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-elasticsearch-bin-local\" (UniqueName: \"kubernetes.io/empty-dir/79b69a1c-270d-4261-b80f-3aff7be1c7bf-elastic-internal-elasticsearch-bin-local\") pod \"elasticsearch-es-default-0\" (UID: \"79b69a1c-270d-4261-b80f-3aff7be1c7bf\") " pod="service-telemetry/elasticsearch-es-default-0" Oct 08 00:17:25 crc kubenswrapper[4967]: I1008 00:17:25.402578 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"downward-api\" (UniqueName: \"kubernetes.io/downward-api/79b69a1c-270d-4261-b80f-3aff7be1c7bf-downward-api\") pod \"elasticsearch-es-default-0\" (UID: \"79b69a1c-270d-4261-b80f-3aff7be1c7bf\") " pod="service-telemetry/elasticsearch-es-default-0" Oct 08 00:17:25 crc kubenswrapper[4967]: I1008 00:17:25.486334 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/elasticsearch-es-default-0" Oct 08 00:17:32 crc kubenswrapper[4967]: E1008 00:17:32.582905 4967 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/cluster-observability-operator/perses-0-1-rhel9-operator@sha256:bfed9f442aea6e8165644f1dc615beea06ec7fd84ea3f8ca393a63d3627c6a7c" Oct 08 00:17:32 crc kubenswrapper[4967]: E1008 00:17:32.583425 4967 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:perses-operator,Image:registry.redhat.io/cluster-observability-operator/perses-0-1-rhel9-operator@sha256:bfed9f442aea6e8165644f1dc615beea06ec7fd84ea3f8ca393a63d3627c6a7c,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:OPERATOR_CONDITION_NAME,Value:cluster-observability-operator.v1.2.2,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{100 -3} {} 100m DecimalSI},memory: {{134217728 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:openshift-service-ca,ReadOnly:true,MountPath:/ca,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6vwgm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000350000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod perses-operator-54bc95c9fb-x476n_openshift-operators(478b5edf-299e-4d19-bc8f-02a5e9774148): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 08 00:17:32 crc kubenswrapper[4967]: E1008 00:17:32.592829 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"perses-operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-operators/perses-operator-54bc95c9fb-x476n" podUID="478b5edf-299e-4d19-bc8f-02a5e9774148" Oct 08 00:17:33 crc kubenswrapper[4967]: E1008 00:17:33.099510 4967 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/amq7/amq-interconnect-operator@sha256:a8b621237c872ded2a1d1d948fbebd693429e4a1ced1d7922406241a078d3d43" Oct 08 00:17:33 crc kubenswrapper[4967]: E1008 00:17:33.099740 4967 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:interconnect-operator,Image:registry.redhat.io/amq7/amq-interconnect-operator@sha256:a8b621237c872ded2a1d1d948fbebd693429e4a1ced1d7922406241a078d3d43,Command:[qdr-operator],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:60000,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:WATCH_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:OPERATOR_NAME,Value:qdr-operator,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_QDROUTERD_IMAGE,Value:registry.redhat.io/amq7/amq-interconnect@sha256:31d87473fa684178a694f9ee331d3c80f2653f9533cb65c2a325752166a077e9,ValueFrom:nil,},EnvVar{Name:OPERATOR_CONDITION_NAME,Value:amq7-interconnect-operator.v1.10.20,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qkpzd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000670000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod interconnect-operator-5bb49f789d-vtczb_service-telemetry(e2154555-42c1-45a2-a30c-70ed7afe09b7): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 08 00:17:33 crc kubenswrapper[4967]: E1008 00:17:33.100956 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"interconnect-operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="service-telemetry/interconnect-operator-5bb49f789d-vtczb" podUID="e2154555-42c1-45a2-a30c-70ed7afe09b7" Oct 08 00:17:33 crc kubenswrapper[4967]: E1008 00:17:33.281284 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"perses-operator\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/cluster-observability-operator/perses-0-1-rhel9-operator@sha256:bfed9f442aea6e8165644f1dc615beea06ec7fd84ea3f8ca393a63d3627c6a7c\\\"\"" pod="openshift-operators/perses-operator-54bc95c9fb-x476n" podUID="478b5edf-299e-4d19-bc8f-02a5e9774148" Oct 08 00:17:33 crc kubenswrapper[4967]: E1008 00:17:33.281298 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"interconnect-operator\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/amq7/amq-interconnect-operator@sha256:a8b621237c872ded2a1d1d948fbebd693429e4a1ced1d7922406241a078d3d43\\\"\"" pod="service-telemetry/interconnect-operator-5bb49f789d-vtczb" podUID="e2154555-42c1-45a2-a30c-70ed7afe09b7" Oct 08 00:17:33 crc kubenswrapper[4967]: I1008 00:17:33.313465 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-96d66748b-w49xk"] Oct 08 00:17:33 crc kubenswrapper[4967]: I1008 00:17:33.332392 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Oct 08 00:17:34 crc kubenswrapper[4967]: I1008 00:17:34.291984 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5fc4456945-q69j8" event={"ID":"a57260c5-8756-489a-be29-6e71c74778bc","Type":"ContainerStarted","Data":"e0e4a9baee03be379b657e2182afbaa13ef52ac8505614d2910782445ae7d848"} Oct 08 00:17:34 crc kubenswrapper[4967]: I1008 00:17:34.296576 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5fc4456945-kpznr" event={"ID":"2a5fd0cb-9c8e-41e9-af18-d18ddf30180e","Type":"ContainerStarted","Data":"433159de5cc676a3f2a98a26152494e5147457a0a0debd9f658f11402eeaf56c"} Oct 08 00:17:34 crc kubenswrapper[4967]: I1008 00:17:34.298353 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"79b69a1c-270d-4261-b80f-3aff7be1c7bf","Type":"ContainerStarted","Data":"c6308dc8cc21baeddc30c7417cdd2bdddd90e29f9c21f3e18891f03150ea6e2c"} Oct 08 00:17:34 crc kubenswrapper[4967]: I1008 00:17:34.301511 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-cc5f78dfc-sgk4j" event={"ID":"d1a621b2-6929-40de-97b1-01a020d70aa2","Type":"ContainerStarted","Data":"527866e7713d66d96e115933ade6d38c594ebdf626a0cb1b452fae2f0c941a0b"} Oct 08 00:17:34 crc kubenswrapper[4967]: I1008 00:17:34.303220 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-cc5f78dfc-sgk4j" Oct 08 00:17:34 crc kubenswrapper[4967]: I1008 00:17:34.304585 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-96d66748b-w49xk" event={"ID":"83068de3-9ba5-4672-b843-b36adba4916e","Type":"ContainerStarted","Data":"a4ee84e32864133ae46cb33c9e38e41dae2f6cd12e3e53a163bd5fdaf111e3da"} Oct 08 00:17:34 crc kubenswrapper[4967]: I1008 00:17:34.305642 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-cc5f78dfc-sgk4j" Oct 08 00:17:34 crc kubenswrapper[4967]: I1008 00:17:34.306738 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-88lmm" event={"ID":"fb355e35-7b9c-4c0d-bb62-aa169f2bef5a","Type":"ContainerStarted","Data":"f512b6c2b9e31894a824ce3ac51ee517c002f70f3e5aaa44b7d984bcbc4be1fa"} Oct 08 00:17:34 crc kubenswrapper[4967]: I1008 00:17:34.311825 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5fc4456945-q69j8" podStartSLOduration=3.080173183 podStartE2EDuration="19.311804673s" podCreationTimestamp="2025-10-08 00:17:15 +0000 UTC" firstStartedPulling="2025-10-08 00:17:16.873238141 +0000 UTC m=+635.523772780" lastFinishedPulling="2025-10-08 00:17:33.104869621 +0000 UTC m=+651.755404270" observedRunningTime="2025-10-08 00:17:34.308452304 +0000 UTC m=+652.958986953" watchObservedRunningTime="2025-10-08 00:17:34.311804673 +0000 UTC m=+652.962339322" Oct 08 00:17:34 crc kubenswrapper[4967]: I1008 00:17:34.362897 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5fc4456945-kpznr" podStartSLOduration=3.295184017 podStartE2EDuration="19.362872445s" podCreationTimestamp="2025-10-08 00:17:15 +0000 UTC" firstStartedPulling="2025-10-08 00:17:16.990788961 +0000 UTC m=+635.641323610" lastFinishedPulling="2025-10-08 00:17:33.058477389 +0000 UTC m=+651.709012038" observedRunningTime="2025-10-08 00:17:34.347542025 +0000 UTC m=+652.998076694" watchObservedRunningTime="2025-10-08 00:17:34.362872445 +0000 UTC m=+653.013407114" Oct 08 00:17:34 crc kubenswrapper[4967]: I1008 00:17:34.397402 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-88lmm" podStartSLOduration=2.881062007 podStartE2EDuration="19.397383168s" podCreationTimestamp="2025-10-08 00:17:15 +0000 UTC" firstStartedPulling="2025-10-08 00:17:16.541916062 +0000 UTC m=+635.192450711" lastFinishedPulling="2025-10-08 00:17:33.058237223 +0000 UTC m=+651.708771872" observedRunningTime="2025-10-08 00:17:34.396443955 +0000 UTC m=+653.046978614" watchObservedRunningTime="2025-10-08 00:17:34.397383168 +0000 UTC m=+653.047917817" Oct 08 00:17:34 crc kubenswrapper[4967]: I1008 00:17:34.435984 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-cc5f78dfc-sgk4j" podStartSLOduration=2.3505885380000002 podStartE2EDuration="18.435966167s" podCreationTimestamp="2025-10-08 00:17:16 +0000 UTC" firstStartedPulling="2025-10-08 00:17:16.978568209 +0000 UTC m=+635.629102858" lastFinishedPulling="2025-10-08 00:17:33.063945828 +0000 UTC m=+651.714480487" observedRunningTime="2025-10-08 00:17:34.42635762 +0000 UTC m=+653.076892269" watchObservedRunningTime="2025-10-08 00:17:34.435966167 +0000 UTC m=+653.086500816" Oct 08 00:17:47 crc kubenswrapper[4967]: I1008 00:17:47.406348 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"79b69a1c-270d-4261-b80f-3aff7be1c7bf","Type":"ContainerStarted","Data":"fadb20ca7d9689851d4f94b90699fd3acdadeaeba57e0232256edbf64fbcf1ea"} Oct 08 00:17:47 crc kubenswrapper[4967]: I1008 00:17:47.408845 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-54bc95c9fb-x476n" event={"ID":"478b5edf-299e-4d19-bc8f-02a5e9774148","Type":"ContainerStarted","Data":"21425ae100bd69d9954cd844d4d025ca0449e7e4742e3d5dbab5bc8f7a692f6f"} Oct 08 00:17:47 crc kubenswrapper[4967]: I1008 00:17:47.409007 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-54bc95c9fb-x476n" Oct 08 00:17:47 crc kubenswrapper[4967]: I1008 00:17:47.410998 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-96d66748b-w49xk" event={"ID":"83068de3-9ba5-4672-b843-b36adba4916e","Type":"ContainerStarted","Data":"ffdb3bf0cac71f5ee88332f2d673e783032f5a46e788568c6329095cff168ef0"} Oct 08 00:17:47 crc kubenswrapper[4967]: I1008 00:17:47.468522 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager-operator/cert-manager-operator-controller-manager-96d66748b-w49xk" podStartSLOduration=10.622265473 podStartE2EDuration="23.468495793s" podCreationTimestamp="2025-10-08 00:17:24 +0000 UTC" firstStartedPulling="2025-10-08 00:17:33.34354589 +0000 UTC m=+651.994080539" lastFinishedPulling="2025-10-08 00:17:46.18977621 +0000 UTC m=+664.840310859" observedRunningTime="2025-10-08 00:17:47.465500702 +0000 UTC m=+666.116035371" watchObservedRunningTime="2025-10-08 00:17:47.468495793 +0000 UTC m=+666.119030452" Oct 08 00:17:47 crc kubenswrapper[4967]: I1008 00:17:47.491091 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-54bc95c9fb-x476n" podStartSLOduration=2.135413342 podStartE2EDuration="31.491068214s" podCreationTimestamp="2025-10-08 00:17:16 +0000 UTC" firstStartedPulling="2025-10-08 00:17:17.127705039 +0000 UTC m=+635.778239688" lastFinishedPulling="2025-10-08 00:17:46.483359911 +0000 UTC m=+665.133894560" observedRunningTime="2025-10-08 00:17:47.487175322 +0000 UTC m=+666.137710031" watchObservedRunningTime="2025-10-08 00:17:47.491068214 +0000 UTC m=+666.141602883" Oct 08 00:17:47 crc kubenswrapper[4967]: I1008 00:17:47.864203 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Oct 08 00:17:47 crc kubenswrapper[4967]: I1008 00:17:47.894626 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Oct 08 00:17:48 crc kubenswrapper[4967]: I1008 00:17:48.418333 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/interconnect-operator-5bb49f789d-vtczb" event={"ID":"e2154555-42c1-45a2-a30c-70ed7afe09b7","Type":"ContainerStarted","Data":"398ab89a5461fe7dfdd9f04b166c4cb1bd8eda9ca9e2905129117021eabadf47"} Oct 08 00:17:49 crc kubenswrapper[4967]: I1008 00:17:49.438725 4967 generic.go:334] "Generic (PLEG): container finished" podID="79b69a1c-270d-4261-b80f-3aff7be1c7bf" containerID="fadb20ca7d9689851d4f94b90699fd3acdadeaeba57e0232256edbf64fbcf1ea" exitCode=0 Oct 08 00:17:49 crc kubenswrapper[4967]: I1008 00:17:49.438789 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"79b69a1c-270d-4261-b80f-3aff7be1c7bf","Type":"ContainerDied","Data":"fadb20ca7d9689851d4f94b90699fd3acdadeaeba57e0232256edbf64fbcf1ea"} Oct 08 00:17:49 crc kubenswrapper[4967]: I1008 00:17:49.498001 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/interconnect-operator-5bb49f789d-vtczb" podStartSLOduration=3.143695348 podStartE2EDuration="37.497981158s" podCreationTimestamp="2025-10-08 00:17:12 +0000 UTC" firstStartedPulling="2025-10-08 00:17:13.179690744 +0000 UTC m=+631.830225413" lastFinishedPulling="2025-10-08 00:17:47.533976574 +0000 UTC m=+666.184511223" observedRunningTime="2025-10-08 00:17:48.439664144 +0000 UTC m=+667.090198783" watchObservedRunningTime="2025-10-08 00:17:49.497981158 +0000 UTC m=+668.148515807" Oct 08 00:17:50 crc kubenswrapper[4967]: I1008 00:17:50.444493 4967 generic.go:334] "Generic (PLEG): container finished" podID="79b69a1c-270d-4261-b80f-3aff7be1c7bf" containerID="69bfe02b150b321957f2b1bdc2b0f629c862ab42040a350c861ee17ed82342d5" exitCode=0 Oct 08 00:17:50 crc kubenswrapper[4967]: I1008 00:17:50.444565 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"79b69a1c-270d-4261-b80f-3aff7be1c7bf","Type":"ContainerDied","Data":"69bfe02b150b321957f2b1bdc2b0f629c862ab42040a350c861ee17ed82342d5"} Oct 08 00:17:50 crc kubenswrapper[4967]: I1008 00:17:50.551849 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-d969966f-mr5n2"] Oct 08 00:17:50 crc kubenswrapper[4967]: I1008 00:17:50.552654 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-d969966f-mr5n2" Oct 08 00:17:50 crc kubenswrapper[4967]: I1008 00:17:50.553939 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Oct 08 00:17:50 crc kubenswrapper[4967]: I1008 00:17:50.554319 4967 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-jnrt8" Oct 08 00:17:50 crc kubenswrapper[4967]: I1008 00:17:50.554736 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Oct 08 00:17:50 crc kubenswrapper[4967]: I1008 00:17:50.564495 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-d969966f-mr5n2"] Oct 08 00:17:50 crc kubenswrapper[4967]: I1008 00:17:50.633746 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e13c83b9-8e80-4bf3-987d-a2b8de2de7be-bound-sa-token\") pod \"cert-manager-webhook-d969966f-mr5n2\" (UID: \"e13c83b9-8e80-4bf3-987d-a2b8de2de7be\") " pod="cert-manager/cert-manager-webhook-d969966f-mr5n2" Oct 08 00:17:50 crc kubenswrapper[4967]: I1008 00:17:50.633851 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2cwth\" (UniqueName: \"kubernetes.io/projected/e13c83b9-8e80-4bf3-987d-a2b8de2de7be-kube-api-access-2cwth\") pod \"cert-manager-webhook-d969966f-mr5n2\" (UID: \"e13c83b9-8e80-4bf3-987d-a2b8de2de7be\") " pod="cert-manager/cert-manager-webhook-d969966f-mr5n2" Oct 08 00:17:50 crc kubenswrapper[4967]: I1008 00:17:50.735170 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2cwth\" (UniqueName: \"kubernetes.io/projected/e13c83b9-8e80-4bf3-987d-a2b8de2de7be-kube-api-access-2cwth\") pod \"cert-manager-webhook-d969966f-mr5n2\" (UID: \"e13c83b9-8e80-4bf3-987d-a2b8de2de7be\") " pod="cert-manager/cert-manager-webhook-d969966f-mr5n2" Oct 08 00:17:50 crc kubenswrapper[4967]: I1008 00:17:50.735238 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e13c83b9-8e80-4bf3-987d-a2b8de2de7be-bound-sa-token\") pod \"cert-manager-webhook-d969966f-mr5n2\" (UID: \"e13c83b9-8e80-4bf3-987d-a2b8de2de7be\") " pod="cert-manager/cert-manager-webhook-d969966f-mr5n2" Oct 08 00:17:50 crc kubenswrapper[4967]: I1008 00:17:50.764434 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e13c83b9-8e80-4bf3-987d-a2b8de2de7be-bound-sa-token\") pod \"cert-manager-webhook-d969966f-mr5n2\" (UID: \"e13c83b9-8e80-4bf3-987d-a2b8de2de7be\") " pod="cert-manager/cert-manager-webhook-d969966f-mr5n2" Oct 08 00:17:50 crc kubenswrapper[4967]: I1008 00:17:50.764615 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2cwth\" (UniqueName: \"kubernetes.io/projected/e13c83b9-8e80-4bf3-987d-a2b8de2de7be-kube-api-access-2cwth\") pod \"cert-manager-webhook-d969966f-mr5n2\" (UID: \"e13c83b9-8e80-4bf3-987d-a2b8de2de7be\") " pod="cert-manager/cert-manager-webhook-d969966f-mr5n2" Oct 08 00:17:50 crc kubenswrapper[4967]: I1008 00:17:50.872612 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-d969966f-mr5n2" Oct 08 00:17:51 crc kubenswrapper[4967]: I1008 00:17:51.341518 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-d969966f-mr5n2"] Oct 08 00:17:51 crc kubenswrapper[4967]: W1008 00:17:51.360824 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode13c83b9_8e80_4bf3_987d_a2b8de2de7be.slice/crio-7d8352899c7f8ec4022c35f8156955d49948bc18950ada51e3a857dcfc09bbf7 WatchSource:0}: Error finding container 7d8352899c7f8ec4022c35f8156955d49948bc18950ada51e3a857dcfc09bbf7: Status 404 returned error can't find the container with id 7d8352899c7f8ec4022c35f8156955d49948bc18950ada51e3a857dcfc09bbf7 Oct 08 00:17:51 crc kubenswrapper[4967]: I1008 00:17:51.450714 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-d969966f-mr5n2" event={"ID":"e13c83b9-8e80-4bf3-987d-a2b8de2de7be","Type":"ContainerStarted","Data":"7d8352899c7f8ec4022c35f8156955d49948bc18950ada51e3a857dcfc09bbf7"} Oct 08 00:17:51 crc kubenswrapper[4967]: I1008 00:17:51.453461 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"79b69a1c-270d-4261-b80f-3aff7be1c7bf","Type":"ContainerStarted","Data":"c6a11a293e71f81ca12a2f50ee613fddbb08b2150c0c3b5e435a5eac800df2db"} Oct 08 00:17:51 crc kubenswrapper[4967]: I1008 00:17:51.453826 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="service-telemetry/elasticsearch-es-default-0" Oct 08 00:17:51 crc kubenswrapper[4967]: I1008 00:17:51.539065 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/elasticsearch-es-default-0" podStartSLOduration=13.289782508 podStartE2EDuration="26.539038207s" podCreationTimestamp="2025-10-08 00:17:25 +0000 UTC" firstStartedPulling="2025-10-08 00:17:33.354561488 +0000 UTC m=+652.005096137" lastFinishedPulling="2025-10-08 00:17:46.603817197 +0000 UTC m=+665.254351836" observedRunningTime="2025-10-08 00:17:51.534456169 +0000 UTC m=+670.184990838" watchObservedRunningTime="2025-10-08 00:17:51.539038207 +0000 UTC m=+670.189572856" Oct 08 00:17:53 crc kubenswrapper[4967]: I1008 00:17:53.610183 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7d9f95dbf-6r7sw"] Oct 08 00:17:53 crc kubenswrapper[4967]: I1008 00:17:53.611291 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7d9f95dbf-6r7sw" Oct 08 00:17:53 crc kubenswrapper[4967]: I1008 00:17:53.613965 4967 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-n5dwg" Oct 08 00:17:53 crc kubenswrapper[4967]: I1008 00:17:53.618164 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7d9f95dbf-6r7sw"] Oct 08 00:17:53 crc kubenswrapper[4967]: I1008 00:17:53.681532 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8b29e193-8366-4f9e-b3c8-761aae502121-bound-sa-token\") pod \"cert-manager-cainjector-7d9f95dbf-6r7sw\" (UID: \"8b29e193-8366-4f9e-b3c8-761aae502121\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-6r7sw" Oct 08 00:17:53 crc kubenswrapper[4967]: I1008 00:17:53.681621 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l8vfq\" (UniqueName: \"kubernetes.io/projected/8b29e193-8366-4f9e-b3c8-761aae502121-kube-api-access-l8vfq\") pod \"cert-manager-cainjector-7d9f95dbf-6r7sw\" (UID: \"8b29e193-8366-4f9e-b3c8-761aae502121\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-6r7sw" Oct 08 00:17:53 crc kubenswrapper[4967]: I1008 00:17:53.782916 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l8vfq\" (UniqueName: \"kubernetes.io/projected/8b29e193-8366-4f9e-b3c8-761aae502121-kube-api-access-l8vfq\") pod \"cert-manager-cainjector-7d9f95dbf-6r7sw\" (UID: \"8b29e193-8366-4f9e-b3c8-761aae502121\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-6r7sw" Oct 08 00:17:53 crc kubenswrapper[4967]: I1008 00:17:53.783021 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8b29e193-8366-4f9e-b3c8-761aae502121-bound-sa-token\") pod \"cert-manager-cainjector-7d9f95dbf-6r7sw\" (UID: \"8b29e193-8366-4f9e-b3c8-761aae502121\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-6r7sw" Oct 08 00:17:53 crc kubenswrapper[4967]: I1008 00:17:53.799054 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8b29e193-8366-4f9e-b3c8-761aae502121-bound-sa-token\") pod \"cert-manager-cainjector-7d9f95dbf-6r7sw\" (UID: \"8b29e193-8366-4f9e-b3c8-761aae502121\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-6r7sw" Oct 08 00:17:53 crc kubenswrapper[4967]: I1008 00:17:53.800627 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l8vfq\" (UniqueName: \"kubernetes.io/projected/8b29e193-8366-4f9e-b3c8-761aae502121-kube-api-access-l8vfq\") pod \"cert-manager-cainjector-7d9f95dbf-6r7sw\" (UID: \"8b29e193-8366-4f9e-b3c8-761aae502121\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-6r7sw" Oct 08 00:17:53 crc kubenswrapper[4967]: I1008 00:17:53.937117 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7d9f95dbf-6r7sw" Oct 08 00:17:55 crc kubenswrapper[4967]: I1008 00:17:55.813797 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7d9f95dbf-6r7sw"] Oct 08 00:17:56 crc kubenswrapper[4967]: I1008 00:17:56.488036 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-d969966f-mr5n2" event={"ID":"e13c83b9-8e80-4bf3-987d-a2b8de2de7be","Type":"ContainerStarted","Data":"6a061ddb1d5b8e565fd4ccf176f56192747b8218fe853eed4d27d9c3eb2603b4"} Oct 08 00:17:56 crc kubenswrapper[4967]: I1008 00:17:56.488971 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-d969966f-mr5n2" Oct 08 00:17:56 crc kubenswrapper[4967]: I1008 00:17:56.490018 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7d9f95dbf-6r7sw" event={"ID":"8b29e193-8366-4f9e-b3c8-761aae502121","Type":"ContainerStarted","Data":"3dd9db5b00b5c11de5660eb0ab7578e4fba2810e8039ead2641af026f21f5da6"} Oct 08 00:17:56 crc kubenswrapper[4967]: I1008 00:17:56.490058 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7d9f95dbf-6r7sw" event={"ID":"8b29e193-8366-4f9e-b3c8-761aae502121","Type":"ContainerStarted","Data":"6b10f07ed342e10a34c5b43f72214a474865ba9c830e5b3337ca936426632bde"} Oct 08 00:17:56 crc kubenswrapper[4967]: I1008 00:17:56.532566 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-d969966f-mr5n2" podStartSLOduration=2.102052751 podStartE2EDuration="6.532537308s" podCreationTimestamp="2025-10-08 00:17:50 +0000 UTC" firstStartedPulling="2025-10-08 00:17:51.362359928 +0000 UTC m=+670.012894567" lastFinishedPulling="2025-10-08 00:17:55.792844465 +0000 UTC m=+674.443379124" observedRunningTime="2025-10-08 00:17:56.52919252 +0000 UTC m=+675.179727179" watchObservedRunningTime="2025-10-08 00:17:56.532537308 +0000 UTC m=+675.183071977" Oct 08 00:17:56 crc kubenswrapper[4967]: I1008 00:17:56.551584 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7d9f95dbf-6r7sw" podStartSLOduration=3.551561376 podStartE2EDuration="3.551561376s" podCreationTimestamp="2025-10-08 00:17:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:17:56.546877686 +0000 UTC m=+675.197412345" watchObservedRunningTime="2025-10-08 00:17:56.551561376 +0000 UTC m=+675.202096025" Oct 08 00:17:56 crc kubenswrapper[4967]: I1008 00:17:56.667633 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-54bc95c9fb-x476n" Oct 08 00:18:00 crc kubenswrapper[4967]: I1008 00:18:00.577350 4967 prober.go:107] "Probe failed" probeType="Readiness" pod="service-telemetry/elasticsearch-es-default-0" podUID="79b69a1c-270d-4261-b80f-3aff7be1c7bf" containerName="elasticsearch" probeResult="failure" output=< Oct 08 00:18:00 crc kubenswrapper[4967]: {"timestamp": "2025-10-08T00:18:00+00:00", "message": "readiness probe failed", "curl_rc": "7"} Oct 08 00:18:00 crc kubenswrapper[4967]: > Oct 08 00:18:00 crc kubenswrapper[4967]: I1008 00:18:00.885297 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-d969966f-mr5n2" Oct 08 00:18:03 crc kubenswrapper[4967]: I1008 00:18:03.966404 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/service-telemetry-operator-1-build"] Oct 08 00:18:03 crc kubenswrapper[4967]: I1008 00:18:03.967776 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-1-build" Oct 08 00:18:03 crc kubenswrapper[4967]: I1008 00:18:03.974451 4967 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-r7snr" Oct 08 00:18:03 crc kubenswrapper[4967]: I1008 00:18:03.974756 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-1-sys-config" Oct 08 00:18:03 crc kubenswrapper[4967]: I1008 00:18:03.975088 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-1-global-ca" Oct 08 00:18:03 crc kubenswrapper[4967]: I1008 00:18:03.975771 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-1-ca" Oct 08 00:18:04 crc kubenswrapper[4967]: I1008 00:18:04.004318 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-1-build"] Oct 08 00:18:04 crc kubenswrapper[4967]: I1008 00:18:04.022212 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/36b9e156-0960-4be2-90b7-8abe8c34d1c7-build-ca-bundles\") pod \"service-telemetry-operator-1-build\" (UID: \"36b9e156-0960-4be2-90b7-8abe8c34d1c7\") " pod="service-telemetry/service-telemetry-operator-1-build" Oct 08 00:18:04 crc kubenswrapper[4967]: I1008 00:18:04.023516 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/36b9e156-0960-4be2-90b7-8abe8c34d1c7-node-pullsecrets\") pod \"service-telemetry-operator-1-build\" (UID: \"36b9e156-0960-4be2-90b7-8abe8c34d1c7\") " pod="service-telemetry/service-telemetry-operator-1-build" Oct 08 00:18:04 crc kubenswrapper[4967]: I1008 00:18:04.023567 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/36b9e156-0960-4be2-90b7-8abe8c34d1c7-buildcachedir\") pod \"service-telemetry-operator-1-build\" (UID: \"36b9e156-0960-4be2-90b7-8abe8c34d1c7\") " pod="service-telemetry/service-telemetry-operator-1-build" Oct 08 00:18:04 crc kubenswrapper[4967]: I1008 00:18:04.023585 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/36b9e156-0960-4be2-90b7-8abe8c34d1c7-build-system-configs\") pod \"service-telemetry-operator-1-build\" (UID: \"36b9e156-0960-4be2-90b7-8abe8c34d1c7\") " pod="service-telemetry/service-telemetry-operator-1-build" Oct 08 00:18:04 crc kubenswrapper[4967]: I1008 00:18:04.023607 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hbg9g\" (UniqueName: \"kubernetes.io/projected/36b9e156-0960-4be2-90b7-8abe8c34d1c7-kube-api-access-hbg9g\") pod \"service-telemetry-operator-1-build\" (UID: \"36b9e156-0960-4be2-90b7-8abe8c34d1c7\") " pod="service-telemetry/service-telemetry-operator-1-build" Oct 08 00:18:04 crc kubenswrapper[4967]: I1008 00:18:04.023626 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/36b9e156-0960-4be2-90b7-8abe8c34d1c7-buildworkdir\") pod \"service-telemetry-operator-1-build\" (UID: \"36b9e156-0960-4be2-90b7-8abe8c34d1c7\") " pod="service-telemetry/service-telemetry-operator-1-build" Oct 08 00:18:04 crc kubenswrapper[4967]: I1008 00:18:04.023646 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/36b9e156-0960-4be2-90b7-8abe8c34d1c7-container-storage-run\") pod \"service-telemetry-operator-1-build\" (UID: \"36b9e156-0960-4be2-90b7-8abe8c34d1c7\") " pod="service-telemetry/service-telemetry-operator-1-build" Oct 08 00:18:04 crc kubenswrapper[4967]: I1008 00:18:04.023690 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-r7snr-push\" (UniqueName: \"kubernetes.io/secret/36b9e156-0960-4be2-90b7-8abe8c34d1c7-builder-dockercfg-r7snr-push\") pod \"service-telemetry-operator-1-build\" (UID: \"36b9e156-0960-4be2-90b7-8abe8c34d1c7\") " pod="service-telemetry/service-telemetry-operator-1-build" Oct 08 00:18:04 crc kubenswrapper[4967]: I1008 00:18:04.023708 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/36b9e156-0960-4be2-90b7-8abe8c34d1c7-build-proxy-ca-bundles\") pod \"service-telemetry-operator-1-build\" (UID: \"36b9e156-0960-4be2-90b7-8abe8c34d1c7\") " pod="service-telemetry/service-telemetry-operator-1-build" Oct 08 00:18:04 crc kubenswrapper[4967]: I1008 00:18:04.023736 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-r7snr-pull\" (UniqueName: \"kubernetes.io/secret/36b9e156-0960-4be2-90b7-8abe8c34d1c7-builder-dockercfg-r7snr-pull\") pod \"service-telemetry-operator-1-build\" (UID: \"36b9e156-0960-4be2-90b7-8abe8c34d1c7\") " pod="service-telemetry/service-telemetry-operator-1-build" Oct 08 00:18:04 crc kubenswrapper[4967]: I1008 00:18:04.023763 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/36b9e156-0960-4be2-90b7-8abe8c34d1c7-container-storage-root\") pod \"service-telemetry-operator-1-build\" (UID: \"36b9e156-0960-4be2-90b7-8abe8c34d1c7\") " pod="service-telemetry/service-telemetry-operator-1-build" Oct 08 00:18:04 crc kubenswrapper[4967]: I1008 00:18:04.023784 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/36b9e156-0960-4be2-90b7-8abe8c34d1c7-build-blob-cache\") pod \"service-telemetry-operator-1-build\" (UID: \"36b9e156-0960-4be2-90b7-8abe8c34d1c7\") " pod="service-telemetry/service-telemetry-operator-1-build" Oct 08 00:18:04 crc kubenswrapper[4967]: I1008 00:18:04.125340 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/36b9e156-0960-4be2-90b7-8abe8c34d1c7-build-ca-bundles\") pod \"service-telemetry-operator-1-build\" (UID: \"36b9e156-0960-4be2-90b7-8abe8c34d1c7\") " pod="service-telemetry/service-telemetry-operator-1-build" Oct 08 00:18:04 crc kubenswrapper[4967]: I1008 00:18:04.125623 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/36b9e156-0960-4be2-90b7-8abe8c34d1c7-node-pullsecrets\") pod \"service-telemetry-operator-1-build\" (UID: \"36b9e156-0960-4be2-90b7-8abe8c34d1c7\") " pod="service-telemetry/service-telemetry-operator-1-build" Oct 08 00:18:04 crc kubenswrapper[4967]: I1008 00:18:04.125798 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/36b9e156-0960-4be2-90b7-8abe8c34d1c7-buildcachedir\") pod \"service-telemetry-operator-1-build\" (UID: \"36b9e156-0960-4be2-90b7-8abe8c34d1c7\") " pod="service-telemetry/service-telemetry-operator-1-build" Oct 08 00:18:04 crc kubenswrapper[4967]: I1008 00:18:04.125917 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/36b9e156-0960-4be2-90b7-8abe8c34d1c7-build-system-configs\") pod \"service-telemetry-operator-1-build\" (UID: \"36b9e156-0960-4be2-90b7-8abe8c34d1c7\") " pod="service-telemetry/service-telemetry-operator-1-build" Oct 08 00:18:04 crc kubenswrapper[4967]: I1008 00:18:04.126050 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hbg9g\" (UniqueName: \"kubernetes.io/projected/36b9e156-0960-4be2-90b7-8abe8c34d1c7-kube-api-access-hbg9g\") pod \"service-telemetry-operator-1-build\" (UID: \"36b9e156-0960-4be2-90b7-8abe8c34d1c7\") " pod="service-telemetry/service-telemetry-operator-1-build" Oct 08 00:18:04 crc kubenswrapper[4967]: I1008 00:18:04.126148 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/36b9e156-0960-4be2-90b7-8abe8c34d1c7-buildworkdir\") pod \"service-telemetry-operator-1-build\" (UID: \"36b9e156-0960-4be2-90b7-8abe8c34d1c7\") " pod="service-telemetry/service-telemetry-operator-1-build" Oct 08 00:18:04 crc kubenswrapper[4967]: I1008 00:18:04.126249 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/36b9e156-0960-4be2-90b7-8abe8c34d1c7-container-storage-run\") pod \"service-telemetry-operator-1-build\" (UID: \"36b9e156-0960-4be2-90b7-8abe8c34d1c7\") " pod="service-telemetry/service-telemetry-operator-1-build" Oct 08 00:18:04 crc kubenswrapper[4967]: I1008 00:18:04.126346 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-r7snr-push\" (UniqueName: \"kubernetes.io/secret/36b9e156-0960-4be2-90b7-8abe8c34d1c7-builder-dockercfg-r7snr-push\") pod \"service-telemetry-operator-1-build\" (UID: \"36b9e156-0960-4be2-90b7-8abe8c34d1c7\") " pod="service-telemetry/service-telemetry-operator-1-build" Oct 08 00:18:04 crc kubenswrapper[4967]: I1008 00:18:04.126450 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/36b9e156-0960-4be2-90b7-8abe8c34d1c7-build-proxy-ca-bundles\") pod \"service-telemetry-operator-1-build\" (UID: \"36b9e156-0960-4be2-90b7-8abe8c34d1c7\") " pod="service-telemetry/service-telemetry-operator-1-build" Oct 08 00:18:04 crc kubenswrapper[4967]: I1008 00:18:04.126579 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-r7snr-pull\" (UniqueName: \"kubernetes.io/secret/36b9e156-0960-4be2-90b7-8abe8c34d1c7-builder-dockercfg-r7snr-pull\") pod \"service-telemetry-operator-1-build\" (UID: \"36b9e156-0960-4be2-90b7-8abe8c34d1c7\") " pod="service-telemetry/service-telemetry-operator-1-build" Oct 08 00:18:04 crc kubenswrapper[4967]: I1008 00:18:04.126721 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/36b9e156-0960-4be2-90b7-8abe8c34d1c7-buildworkdir\") pod \"service-telemetry-operator-1-build\" (UID: \"36b9e156-0960-4be2-90b7-8abe8c34d1c7\") " pod="service-telemetry/service-telemetry-operator-1-build" Oct 08 00:18:04 crc kubenswrapper[4967]: I1008 00:18:04.126646 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/36b9e156-0960-4be2-90b7-8abe8c34d1c7-build-system-configs\") pod \"service-telemetry-operator-1-build\" (UID: \"36b9e156-0960-4be2-90b7-8abe8c34d1c7\") " pod="service-telemetry/service-telemetry-operator-1-build" Oct 08 00:18:04 crc kubenswrapper[4967]: I1008 00:18:04.125803 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/36b9e156-0960-4be2-90b7-8abe8c34d1c7-node-pullsecrets\") pod \"service-telemetry-operator-1-build\" (UID: \"36b9e156-0960-4be2-90b7-8abe8c34d1c7\") " pod="service-telemetry/service-telemetry-operator-1-build" Oct 08 00:18:04 crc kubenswrapper[4967]: I1008 00:18:04.126509 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/36b9e156-0960-4be2-90b7-8abe8c34d1c7-build-ca-bundles\") pod \"service-telemetry-operator-1-build\" (UID: \"36b9e156-0960-4be2-90b7-8abe8c34d1c7\") " pod="service-telemetry/service-telemetry-operator-1-build" Oct 08 00:18:04 crc kubenswrapper[4967]: I1008 00:18:04.125836 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/36b9e156-0960-4be2-90b7-8abe8c34d1c7-buildcachedir\") pod \"service-telemetry-operator-1-build\" (UID: \"36b9e156-0960-4be2-90b7-8abe8c34d1c7\") " pod="service-telemetry/service-telemetry-operator-1-build" Oct 08 00:18:04 crc kubenswrapper[4967]: I1008 00:18:04.126871 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/36b9e156-0960-4be2-90b7-8abe8c34d1c7-container-storage-run\") pod \"service-telemetry-operator-1-build\" (UID: \"36b9e156-0960-4be2-90b7-8abe8c34d1c7\") " pod="service-telemetry/service-telemetry-operator-1-build" Oct 08 00:18:04 crc kubenswrapper[4967]: I1008 00:18:04.127046 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/36b9e156-0960-4be2-90b7-8abe8c34d1c7-container-storage-root\") pod \"service-telemetry-operator-1-build\" (UID: \"36b9e156-0960-4be2-90b7-8abe8c34d1c7\") " pod="service-telemetry/service-telemetry-operator-1-build" Oct 08 00:18:04 crc kubenswrapper[4967]: I1008 00:18:04.127159 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/36b9e156-0960-4be2-90b7-8abe8c34d1c7-build-blob-cache\") pod \"service-telemetry-operator-1-build\" (UID: \"36b9e156-0960-4be2-90b7-8abe8c34d1c7\") " pod="service-telemetry/service-telemetry-operator-1-build" Oct 08 00:18:04 crc kubenswrapper[4967]: I1008 00:18:04.127243 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/36b9e156-0960-4be2-90b7-8abe8c34d1c7-container-storage-root\") pod \"service-telemetry-operator-1-build\" (UID: \"36b9e156-0960-4be2-90b7-8abe8c34d1c7\") " pod="service-telemetry/service-telemetry-operator-1-build" Oct 08 00:18:04 crc kubenswrapper[4967]: I1008 00:18:04.127433 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/36b9e156-0960-4be2-90b7-8abe8c34d1c7-build-proxy-ca-bundles\") pod \"service-telemetry-operator-1-build\" (UID: \"36b9e156-0960-4be2-90b7-8abe8c34d1c7\") " pod="service-telemetry/service-telemetry-operator-1-build" Oct 08 00:18:04 crc kubenswrapper[4967]: I1008 00:18:04.127797 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/36b9e156-0960-4be2-90b7-8abe8c34d1c7-build-blob-cache\") pod \"service-telemetry-operator-1-build\" (UID: \"36b9e156-0960-4be2-90b7-8abe8c34d1c7\") " pod="service-telemetry/service-telemetry-operator-1-build" Oct 08 00:18:04 crc kubenswrapper[4967]: I1008 00:18:04.132013 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-r7snr-pull\" (UniqueName: \"kubernetes.io/secret/36b9e156-0960-4be2-90b7-8abe8c34d1c7-builder-dockercfg-r7snr-pull\") pod \"service-telemetry-operator-1-build\" (UID: \"36b9e156-0960-4be2-90b7-8abe8c34d1c7\") " pod="service-telemetry/service-telemetry-operator-1-build" Oct 08 00:18:04 crc kubenswrapper[4967]: I1008 00:18:04.132760 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-r7snr-push\" (UniqueName: \"kubernetes.io/secret/36b9e156-0960-4be2-90b7-8abe8c34d1c7-builder-dockercfg-r7snr-push\") pod \"service-telemetry-operator-1-build\" (UID: \"36b9e156-0960-4be2-90b7-8abe8c34d1c7\") " pod="service-telemetry/service-telemetry-operator-1-build" Oct 08 00:18:04 crc kubenswrapper[4967]: I1008 00:18:04.161511 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hbg9g\" (UniqueName: \"kubernetes.io/projected/36b9e156-0960-4be2-90b7-8abe8c34d1c7-kube-api-access-hbg9g\") pod \"service-telemetry-operator-1-build\" (UID: \"36b9e156-0960-4be2-90b7-8abe8c34d1c7\") " pod="service-telemetry/service-telemetry-operator-1-build" Oct 08 00:18:04 crc kubenswrapper[4967]: I1008 00:18:04.284326 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-1-build" Oct 08 00:18:04 crc kubenswrapper[4967]: I1008 00:18:04.547164 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-1-build"] Oct 08 00:18:04 crc kubenswrapper[4967]: W1008 00:18:04.558732 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod36b9e156_0960_4be2_90b7_8abe8c34d1c7.slice/crio-0b69261a4c9913ce438892715d5fb83e32006cb986365f731de58c1e1bbc78d4 WatchSource:0}: Error finding container 0b69261a4c9913ce438892715d5fb83e32006cb986365f731de58c1e1bbc78d4: Status 404 returned error can't find the container with id 0b69261a4c9913ce438892715d5fb83e32006cb986365f731de58c1e1bbc78d4 Oct 08 00:18:05 crc kubenswrapper[4967]: I1008 00:18:05.557942 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-1-build" event={"ID":"36b9e156-0960-4be2-90b7-8abe8c34d1c7","Type":"ContainerStarted","Data":"0b69261a4c9913ce438892715d5fb83e32006cb986365f731de58c1e1bbc78d4"} Oct 08 00:18:05 crc kubenswrapper[4967]: I1008 00:18:05.754545 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="service-telemetry/elasticsearch-es-default-0" Oct 08 00:18:10 crc kubenswrapper[4967]: I1008 00:18:10.865138 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-7d4cc89fcb-fftc7"] Oct 08 00:18:10 crc kubenswrapper[4967]: I1008 00:18:10.866287 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-7d4cc89fcb-fftc7" Oct 08 00:18:10 crc kubenswrapper[4967]: I1008 00:18:10.868905 4967 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-5557d" Oct 08 00:18:10 crc kubenswrapper[4967]: I1008 00:18:10.879743 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-7d4cc89fcb-fftc7"] Oct 08 00:18:10 crc kubenswrapper[4967]: I1008 00:18:10.921778 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/59077174-55c7-4ca3-8ad4-9252240ead8c-bound-sa-token\") pod \"cert-manager-7d4cc89fcb-fftc7\" (UID: \"59077174-55c7-4ca3-8ad4-9252240ead8c\") " pod="cert-manager/cert-manager-7d4cc89fcb-fftc7" Oct 08 00:18:10 crc kubenswrapper[4967]: I1008 00:18:10.921857 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gs4cz\" (UniqueName: \"kubernetes.io/projected/59077174-55c7-4ca3-8ad4-9252240ead8c-kube-api-access-gs4cz\") pod \"cert-manager-7d4cc89fcb-fftc7\" (UID: \"59077174-55c7-4ca3-8ad4-9252240ead8c\") " pod="cert-manager/cert-manager-7d4cc89fcb-fftc7" Oct 08 00:18:11 crc kubenswrapper[4967]: I1008 00:18:11.023480 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/59077174-55c7-4ca3-8ad4-9252240ead8c-bound-sa-token\") pod \"cert-manager-7d4cc89fcb-fftc7\" (UID: \"59077174-55c7-4ca3-8ad4-9252240ead8c\") " pod="cert-manager/cert-manager-7d4cc89fcb-fftc7" Oct 08 00:18:11 crc kubenswrapper[4967]: I1008 00:18:11.023556 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gs4cz\" (UniqueName: \"kubernetes.io/projected/59077174-55c7-4ca3-8ad4-9252240ead8c-kube-api-access-gs4cz\") pod \"cert-manager-7d4cc89fcb-fftc7\" (UID: \"59077174-55c7-4ca3-8ad4-9252240ead8c\") " pod="cert-manager/cert-manager-7d4cc89fcb-fftc7" Oct 08 00:18:11 crc kubenswrapper[4967]: I1008 00:18:11.045084 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/59077174-55c7-4ca3-8ad4-9252240ead8c-bound-sa-token\") pod \"cert-manager-7d4cc89fcb-fftc7\" (UID: \"59077174-55c7-4ca3-8ad4-9252240ead8c\") " pod="cert-manager/cert-manager-7d4cc89fcb-fftc7" Oct 08 00:18:11 crc kubenswrapper[4967]: I1008 00:18:11.045238 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gs4cz\" (UniqueName: \"kubernetes.io/projected/59077174-55c7-4ca3-8ad4-9252240ead8c-kube-api-access-gs4cz\") pod \"cert-manager-7d4cc89fcb-fftc7\" (UID: \"59077174-55c7-4ca3-8ad4-9252240ead8c\") " pod="cert-manager/cert-manager-7d4cc89fcb-fftc7" Oct 08 00:18:11 crc kubenswrapper[4967]: I1008 00:18:11.184953 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-7d4cc89fcb-fftc7" Oct 08 00:18:12 crc kubenswrapper[4967]: I1008 00:18:12.025869 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-7d4cc89fcb-fftc7"] Oct 08 00:18:12 crc kubenswrapper[4967]: I1008 00:18:12.611932 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-7d4cc89fcb-fftc7" event={"ID":"59077174-55c7-4ca3-8ad4-9252240ead8c","Type":"ContainerStarted","Data":"4ee882041b112e7e38fa60575b9f33de949045c758fbb01a0f0d3b773682d39c"} Oct 08 00:18:12 crc kubenswrapper[4967]: I1008 00:18:12.612189 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-7d4cc89fcb-fftc7" event={"ID":"59077174-55c7-4ca3-8ad4-9252240ead8c","Type":"ContainerStarted","Data":"ed218991da0a86d614f7937c0f23c307111add464fef6a2a2275c0268b7e4e85"} Oct 08 00:18:12 crc kubenswrapper[4967]: I1008 00:18:12.613636 4967 generic.go:334] "Generic (PLEG): container finished" podID="36b9e156-0960-4be2-90b7-8abe8c34d1c7" containerID="dd5311ce8b169d403504f774c3492781abc4472d38e9278c261868c44077be83" exitCode=0 Oct 08 00:18:12 crc kubenswrapper[4967]: I1008 00:18:12.613697 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-1-build" event={"ID":"36b9e156-0960-4be2-90b7-8abe8c34d1c7","Type":"ContainerDied","Data":"dd5311ce8b169d403504f774c3492781abc4472d38e9278c261868c44077be83"} Oct 08 00:18:12 crc kubenswrapper[4967]: I1008 00:18:12.633662 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-7d4cc89fcb-fftc7" podStartSLOduration=2.633638124 podStartE2EDuration="2.633638124s" podCreationTimestamp="2025-10-08 00:18:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:18:12.631272628 +0000 UTC m=+691.281807287" watchObservedRunningTime="2025-10-08 00:18:12.633638124 +0000 UTC m=+691.284172773" Oct 08 00:18:13 crc kubenswrapper[4967]: I1008 00:18:13.632074 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-1-build" event={"ID":"36b9e156-0960-4be2-90b7-8abe8c34d1c7","Type":"ContainerStarted","Data":"59b34e9e905cad4da9574fe662b46abeed4961716c6db3bf78dceab52c3b0d9c"} Oct 08 00:18:13 crc kubenswrapper[4967]: I1008 00:18:13.670734 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/service-telemetry-operator-1-build" podStartSLOduration=3.360393665 podStartE2EDuration="10.670709207s" podCreationTimestamp="2025-10-08 00:18:03 +0000 UTC" firstStartedPulling="2025-10-08 00:18:04.560428143 +0000 UTC m=+683.210962792" lastFinishedPulling="2025-10-08 00:18:11.870743685 +0000 UTC m=+690.521278334" observedRunningTime="2025-10-08 00:18:13.665614737 +0000 UTC m=+692.316149406" watchObservedRunningTime="2025-10-08 00:18:13.670709207 +0000 UTC m=+692.321243856" Oct 08 00:18:14 crc kubenswrapper[4967]: I1008 00:18:14.199145 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/service-telemetry-operator-1-build"] Oct 08 00:18:15 crc kubenswrapper[4967]: I1008 00:18:15.646702 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="service-telemetry/service-telemetry-operator-1-build" podUID="36b9e156-0960-4be2-90b7-8abe8c34d1c7" containerName="docker-build" containerID="cri-o://59b34e9e905cad4da9574fe662b46abeed4961716c6db3bf78dceab52c3b0d9c" gracePeriod=30 Oct 08 00:18:15 crc kubenswrapper[4967]: I1008 00:18:15.790913 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/service-telemetry-operator-2-build"] Oct 08 00:18:15 crc kubenswrapper[4967]: I1008 00:18:15.792606 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-2-build" Oct 08 00:18:15 crc kubenswrapper[4967]: I1008 00:18:15.795531 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-2-ca" Oct 08 00:18:15 crc kubenswrapper[4967]: I1008 00:18:15.795611 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-2-global-ca" Oct 08 00:18:15 crc kubenswrapper[4967]: I1008 00:18:15.795539 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-2-sys-config" Oct 08 00:18:15 crc kubenswrapper[4967]: I1008 00:18:15.828307 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/76078656-a41c-4e34-ad5d-ca11ebaa3ff3-build-proxy-ca-bundles\") pod \"service-telemetry-operator-2-build\" (UID: \"76078656-a41c-4e34-ad5d-ca11ebaa3ff3\") " pod="service-telemetry/service-telemetry-operator-2-build" Oct 08 00:18:15 crc kubenswrapper[4967]: I1008 00:18:15.828358 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/76078656-a41c-4e34-ad5d-ca11ebaa3ff3-build-blob-cache\") pod \"service-telemetry-operator-2-build\" (UID: \"76078656-a41c-4e34-ad5d-ca11ebaa3ff3\") " pod="service-telemetry/service-telemetry-operator-2-build" Oct 08 00:18:15 crc kubenswrapper[4967]: I1008 00:18:15.828395 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-r7snr-push\" (UniqueName: \"kubernetes.io/secret/76078656-a41c-4e34-ad5d-ca11ebaa3ff3-builder-dockercfg-r7snr-push\") pod \"service-telemetry-operator-2-build\" (UID: \"76078656-a41c-4e34-ad5d-ca11ebaa3ff3\") " pod="service-telemetry/service-telemetry-operator-2-build" Oct 08 00:18:15 crc kubenswrapper[4967]: I1008 00:18:15.828421 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/76078656-a41c-4e34-ad5d-ca11ebaa3ff3-buildcachedir\") pod \"service-telemetry-operator-2-build\" (UID: \"76078656-a41c-4e34-ad5d-ca11ebaa3ff3\") " pod="service-telemetry/service-telemetry-operator-2-build" Oct 08 00:18:15 crc kubenswrapper[4967]: I1008 00:18:15.828477 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/76078656-a41c-4e34-ad5d-ca11ebaa3ff3-build-ca-bundles\") pod \"service-telemetry-operator-2-build\" (UID: \"76078656-a41c-4e34-ad5d-ca11ebaa3ff3\") " pod="service-telemetry/service-telemetry-operator-2-build" Oct 08 00:18:15 crc kubenswrapper[4967]: I1008 00:18:15.828513 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bqgnl\" (UniqueName: \"kubernetes.io/projected/76078656-a41c-4e34-ad5d-ca11ebaa3ff3-kube-api-access-bqgnl\") pod \"service-telemetry-operator-2-build\" (UID: \"76078656-a41c-4e34-ad5d-ca11ebaa3ff3\") " pod="service-telemetry/service-telemetry-operator-2-build" Oct 08 00:18:15 crc kubenswrapper[4967]: I1008 00:18:15.828533 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/76078656-a41c-4e34-ad5d-ca11ebaa3ff3-container-storage-root\") pod \"service-telemetry-operator-2-build\" (UID: \"76078656-a41c-4e34-ad5d-ca11ebaa3ff3\") " pod="service-telemetry/service-telemetry-operator-2-build" Oct 08 00:18:15 crc kubenswrapper[4967]: I1008 00:18:15.828594 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/76078656-a41c-4e34-ad5d-ca11ebaa3ff3-container-storage-run\") pod \"service-telemetry-operator-2-build\" (UID: \"76078656-a41c-4e34-ad5d-ca11ebaa3ff3\") " pod="service-telemetry/service-telemetry-operator-2-build" Oct 08 00:18:15 crc kubenswrapper[4967]: I1008 00:18:15.828619 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/76078656-a41c-4e34-ad5d-ca11ebaa3ff3-buildworkdir\") pod \"service-telemetry-operator-2-build\" (UID: \"76078656-a41c-4e34-ad5d-ca11ebaa3ff3\") " pod="service-telemetry/service-telemetry-operator-2-build" Oct 08 00:18:15 crc kubenswrapper[4967]: I1008 00:18:15.828642 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-r7snr-pull\" (UniqueName: \"kubernetes.io/secret/76078656-a41c-4e34-ad5d-ca11ebaa3ff3-builder-dockercfg-r7snr-pull\") pod \"service-telemetry-operator-2-build\" (UID: \"76078656-a41c-4e34-ad5d-ca11ebaa3ff3\") " pod="service-telemetry/service-telemetry-operator-2-build" Oct 08 00:18:15 crc kubenswrapper[4967]: I1008 00:18:15.828718 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/76078656-a41c-4e34-ad5d-ca11ebaa3ff3-node-pullsecrets\") pod \"service-telemetry-operator-2-build\" (UID: \"76078656-a41c-4e34-ad5d-ca11ebaa3ff3\") " pod="service-telemetry/service-telemetry-operator-2-build" Oct 08 00:18:15 crc kubenswrapper[4967]: I1008 00:18:15.828794 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/76078656-a41c-4e34-ad5d-ca11ebaa3ff3-build-system-configs\") pod \"service-telemetry-operator-2-build\" (UID: \"76078656-a41c-4e34-ad5d-ca11ebaa3ff3\") " pod="service-telemetry/service-telemetry-operator-2-build" Oct 08 00:18:15 crc kubenswrapper[4967]: I1008 00:18:15.834290 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-2-build"] Oct 08 00:18:15 crc kubenswrapper[4967]: I1008 00:18:15.929568 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-r7snr-push\" (UniqueName: \"kubernetes.io/secret/76078656-a41c-4e34-ad5d-ca11ebaa3ff3-builder-dockercfg-r7snr-push\") pod \"service-telemetry-operator-2-build\" (UID: \"76078656-a41c-4e34-ad5d-ca11ebaa3ff3\") " pod="service-telemetry/service-telemetry-operator-2-build" Oct 08 00:18:15 crc kubenswrapper[4967]: I1008 00:18:15.929608 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/76078656-a41c-4e34-ad5d-ca11ebaa3ff3-buildcachedir\") pod \"service-telemetry-operator-2-build\" (UID: \"76078656-a41c-4e34-ad5d-ca11ebaa3ff3\") " pod="service-telemetry/service-telemetry-operator-2-build" Oct 08 00:18:15 crc kubenswrapper[4967]: I1008 00:18:15.929633 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/76078656-a41c-4e34-ad5d-ca11ebaa3ff3-build-ca-bundles\") pod \"service-telemetry-operator-2-build\" (UID: \"76078656-a41c-4e34-ad5d-ca11ebaa3ff3\") " pod="service-telemetry/service-telemetry-operator-2-build" Oct 08 00:18:15 crc kubenswrapper[4967]: I1008 00:18:15.929663 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bqgnl\" (UniqueName: \"kubernetes.io/projected/76078656-a41c-4e34-ad5d-ca11ebaa3ff3-kube-api-access-bqgnl\") pod \"service-telemetry-operator-2-build\" (UID: \"76078656-a41c-4e34-ad5d-ca11ebaa3ff3\") " pod="service-telemetry/service-telemetry-operator-2-build" Oct 08 00:18:15 crc kubenswrapper[4967]: I1008 00:18:15.929722 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/76078656-a41c-4e34-ad5d-ca11ebaa3ff3-container-storage-root\") pod \"service-telemetry-operator-2-build\" (UID: \"76078656-a41c-4e34-ad5d-ca11ebaa3ff3\") " pod="service-telemetry/service-telemetry-operator-2-build" Oct 08 00:18:15 crc kubenswrapper[4967]: I1008 00:18:15.929741 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/76078656-a41c-4e34-ad5d-ca11ebaa3ff3-buildworkdir\") pod \"service-telemetry-operator-2-build\" (UID: \"76078656-a41c-4e34-ad5d-ca11ebaa3ff3\") " pod="service-telemetry/service-telemetry-operator-2-build" Oct 08 00:18:15 crc kubenswrapper[4967]: I1008 00:18:15.929746 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/76078656-a41c-4e34-ad5d-ca11ebaa3ff3-buildcachedir\") pod \"service-telemetry-operator-2-build\" (UID: \"76078656-a41c-4e34-ad5d-ca11ebaa3ff3\") " pod="service-telemetry/service-telemetry-operator-2-build" Oct 08 00:18:15 crc kubenswrapper[4967]: I1008 00:18:15.929755 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/76078656-a41c-4e34-ad5d-ca11ebaa3ff3-container-storage-run\") pod \"service-telemetry-operator-2-build\" (UID: \"76078656-a41c-4e34-ad5d-ca11ebaa3ff3\") " pod="service-telemetry/service-telemetry-operator-2-build" Oct 08 00:18:15 crc kubenswrapper[4967]: I1008 00:18:15.929795 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-r7snr-pull\" (UniqueName: \"kubernetes.io/secret/76078656-a41c-4e34-ad5d-ca11ebaa3ff3-builder-dockercfg-r7snr-pull\") pod \"service-telemetry-operator-2-build\" (UID: \"76078656-a41c-4e34-ad5d-ca11ebaa3ff3\") " pod="service-telemetry/service-telemetry-operator-2-build" Oct 08 00:18:15 crc kubenswrapper[4967]: I1008 00:18:15.929814 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/76078656-a41c-4e34-ad5d-ca11ebaa3ff3-node-pullsecrets\") pod \"service-telemetry-operator-2-build\" (UID: \"76078656-a41c-4e34-ad5d-ca11ebaa3ff3\") " pod="service-telemetry/service-telemetry-operator-2-build" Oct 08 00:18:15 crc kubenswrapper[4967]: I1008 00:18:15.929840 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/76078656-a41c-4e34-ad5d-ca11ebaa3ff3-build-system-configs\") pod \"service-telemetry-operator-2-build\" (UID: \"76078656-a41c-4e34-ad5d-ca11ebaa3ff3\") " pod="service-telemetry/service-telemetry-operator-2-build" Oct 08 00:18:15 crc kubenswrapper[4967]: I1008 00:18:15.929881 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/76078656-a41c-4e34-ad5d-ca11ebaa3ff3-build-proxy-ca-bundles\") pod \"service-telemetry-operator-2-build\" (UID: \"76078656-a41c-4e34-ad5d-ca11ebaa3ff3\") " pod="service-telemetry/service-telemetry-operator-2-build" Oct 08 00:18:15 crc kubenswrapper[4967]: I1008 00:18:15.929898 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/76078656-a41c-4e34-ad5d-ca11ebaa3ff3-build-blob-cache\") pod \"service-telemetry-operator-2-build\" (UID: \"76078656-a41c-4e34-ad5d-ca11ebaa3ff3\") " pod="service-telemetry/service-telemetry-operator-2-build" Oct 08 00:18:15 crc kubenswrapper[4967]: I1008 00:18:15.930219 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/76078656-a41c-4e34-ad5d-ca11ebaa3ff3-container-storage-run\") pod \"service-telemetry-operator-2-build\" (UID: \"76078656-a41c-4e34-ad5d-ca11ebaa3ff3\") " pod="service-telemetry/service-telemetry-operator-2-build" Oct 08 00:18:15 crc kubenswrapper[4967]: I1008 00:18:15.930272 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/76078656-a41c-4e34-ad5d-ca11ebaa3ff3-build-blob-cache\") pod \"service-telemetry-operator-2-build\" (UID: \"76078656-a41c-4e34-ad5d-ca11ebaa3ff3\") " pod="service-telemetry/service-telemetry-operator-2-build" Oct 08 00:18:15 crc kubenswrapper[4967]: I1008 00:18:15.930451 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/76078656-a41c-4e34-ad5d-ca11ebaa3ff3-container-storage-root\") pod \"service-telemetry-operator-2-build\" (UID: \"76078656-a41c-4e34-ad5d-ca11ebaa3ff3\") " pod="service-telemetry/service-telemetry-operator-2-build" Oct 08 00:18:15 crc kubenswrapper[4967]: I1008 00:18:15.930694 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/76078656-a41c-4e34-ad5d-ca11ebaa3ff3-buildworkdir\") pod \"service-telemetry-operator-2-build\" (UID: \"76078656-a41c-4e34-ad5d-ca11ebaa3ff3\") " pod="service-telemetry/service-telemetry-operator-2-build" Oct 08 00:18:15 crc kubenswrapper[4967]: I1008 00:18:15.930725 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/76078656-a41c-4e34-ad5d-ca11ebaa3ff3-node-pullsecrets\") pod \"service-telemetry-operator-2-build\" (UID: \"76078656-a41c-4e34-ad5d-ca11ebaa3ff3\") " pod="service-telemetry/service-telemetry-operator-2-build" Oct 08 00:18:15 crc kubenswrapper[4967]: I1008 00:18:15.930960 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/76078656-a41c-4e34-ad5d-ca11ebaa3ff3-build-ca-bundles\") pod \"service-telemetry-operator-2-build\" (UID: \"76078656-a41c-4e34-ad5d-ca11ebaa3ff3\") " pod="service-telemetry/service-telemetry-operator-2-build" Oct 08 00:18:15 crc kubenswrapper[4967]: I1008 00:18:15.932004 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/76078656-a41c-4e34-ad5d-ca11ebaa3ff3-build-proxy-ca-bundles\") pod \"service-telemetry-operator-2-build\" (UID: \"76078656-a41c-4e34-ad5d-ca11ebaa3ff3\") " pod="service-telemetry/service-telemetry-operator-2-build" Oct 08 00:18:15 crc kubenswrapper[4967]: I1008 00:18:15.932411 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/76078656-a41c-4e34-ad5d-ca11ebaa3ff3-build-system-configs\") pod \"service-telemetry-operator-2-build\" (UID: \"76078656-a41c-4e34-ad5d-ca11ebaa3ff3\") " pod="service-telemetry/service-telemetry-operator-2-build" Oct 08 00:18:15 crc kubenswrapper[4967]: I1008 00:18:15.939808 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-r7snr-push\" (UniqueName: \"kubernetes.io/secret/76078656-a41c-4e34-ad5d-ca11ebaa3ff3-builder-dockercfg-r7snr-push\") pod \"service-telemetry-operator-2-build\" (UID: \"76078656-a41c-4e34-ad5d-ca11ebaa3ff3\") " pod="service-telemetry/service-telemetry-operator-2-build" Oct 08 00:18:15 crc kubenswrapper[4967]: I1008 00:18:15.943897 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-r7snr-pull\" (UniqueName: \"kubernetes.io/secret/76078656-a41c-4e34-ad5d-ca11ebaa3ff3-builder-dockercfg-r7snr-pull\") pod \"service-telemetry-operator-2-build\" (UID: \"76078656-a41c-4e34-ad5d-ca11ebaa3ff3\") " pod="service-telemetry/service-telemetry-operator-2-build" Oct 08 00:18:15 crc kubenswrapper[4967]: I1008 00:18:15.959088 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bqgnl\" (UniqueName: \"kubernetes.io/projected/76078656-a41c-4e34-ad5d-ca11ebaa3ff3-kube-api-access-bqgnl\") pod \"service-telemetry-operator-2-build\" (UID: \"76078656-a41c-4e34-ad5d-ca11ebaa3ff3\") " pod="service-telemetry/service-telemetry-operator-2-build" Oct 08 00:18:16 crc kubenswrapper[4967]: I1008 00:18:16.125386 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-2-build" Oct 08 00:18:17 crc kubenswrapper[4967]: I1008 00:18:17.565719 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-2-build"] Oct 08 00:18:17 crc kubenswrapper[4967]: W1008 00:18:17.572793 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod76078656_a41c_4e34_ad5d_ca11ebaa3ff3.slice/crio-ea9a167c36c73916cee3e8f90dca201b4778ac87941ae7f63d63f3821fd84fe2 WatchSource:0}: Error finding container ea9a167c36c73916cee3e8f90dca201b4778ac87941ae7f63d63f3821fd84fe2: Status 404 returned error can't find the container with id ea9a167c36c73916cee3e8f90dca201b4778ac87941ae7f63d63f3821fd84fe2 Oct 08 00:18:17 crc kubenswrapper[4967]: I1008 00:18:17.662747 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-2-build" event={"ID":"76078656-a41c-4e34-ad5d-ca11ebaa3ff3","Type":"ContainerStarted","Data":"ea9a167c36c73916cee3e8f90dca201b4778ac87941ae7f63d63f3821fd84fe2"} Oct 08 00:18:21 crc kubenswrapper[4967]: I1008 00:18:21.522478 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-1-build_36b9e156-0960-4be2-90b7-8abe8c34d1c7/docker-build/0.log" Oct 08 00:18:21 crc kubenswrapper[4967]: I1008 00:18:21.523579 4967 generic.go:334] "Generic (PLEG): container finished" podID="36b9e156-0960-4be2-90b7-8abe8c34d1c7" containerID="59b34e9e905cad4da9574fe662b46abeed4961716c6db3bf78dceab52c3b0d9c" exitCode=1 Oct 08 00:18:21 crc kubenswrapper[4967]: I1008 00:18:21.523628 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-1-build" event={"ID":"36b9e156-0960-4be2-90b7-8abe8c34d1c7","Type":"ContainerDied","Data":"59b34e9e905cad4da9574fe662b46abeed4961716c6db3bf78dceab52c3b0d9c"} Oct 08 00:18:21 crc kubenswrapper[4967]: I1008 00:18:21.567855 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-1-build_36b9e156-0960-4be2-90b7-8abe8c34d1c7/docker-build/0.log" Oct 08 00:18:21 crc kubenswrapper[4967]: I1008 00:18:21.568447 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-1-build" Oct 08 00:18:21 crc kubenswrapper[4967]: I1008 00:18:21.731347 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-r7snr-push\" (UniqueName: \"kubernetes.io/secret/36b9e156-0960-4be2-90b7-8abe8c34d1c7-builder-dockercfg-r7snr-push\") pod \"36b9e156-0960-4be2-90b7-8abe8c34d1c7\" (UID: \"36b9e156-0960-4be2-90b7-8abe8c34d1c7\") " Oct 08 00:18:21 crc kubenswrapper[4967]: I1008 00:18:21.731429 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/36b9e156-0960-4be2-90b7-8abe8c34d1c7-buildcachedir\") pod \"36b9e156-0960-4be2-90b7-8abe8c34d1c7\" (UID: \"36b9e156-0960-4be2-90b7-8abe8c34d1c7\") " Oct 08 00:18:21 crc kubenswrapper[4967]: I1008 00:18:21.731503 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/36b9e156-0960-4be2-90b7-8abe8c34d1c7-build-proxy-ca-bundles\") pod \"36b9e156-0960-4be2-90b7-8abe8c34d1c7\" (UID: \"36b9e156-0960-4be2-90b7-8abe8c34d1c7\") " Oct 08 00:18:21 crc kubenswrapper[4967]: I1008 00:18:21.731543 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hbg9g\" (UniqueName: \"kubernetes.io/projected/36b9e156-0960-4be2-90b7-8abe8c34d1c7-kube-api-access-hbg9g\") pod \"36b9e156-0960-4be2-90b7-8abe8c34d1c7\" (UID: \"36b9e156-0960-4be2-90b7-8abe8c34d1c7\") " Oct 08 00:18:21 crc kubenswrapper[4967]: I1008 00:18:21.731583 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/36b9e156-0960-4be2-90b7-8abe8c34d1c7-container-storage-run\") pod \"36b9e156-0960-4be2-90b7-8abe8c34d1c7\" (UID: \"36b9e156-0960-4be2-90b7-8abe8c34d1c7\") " Oct 08 00:18:21 crc kubenswrapper[4967]: I1008 00:18:21.731633 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/36b9e156-0960-4be2-90b7-8abe8c34d1c7-build-ca-bundles\") pod \"36b9e156-0960-4be2-90b7-8abe8c34d1c7\" (UID: \"36b9e156-0960-4be2-90b7-8abe8c34d1c7\") " Oct 08 00:18:21 crc kubenswrapper[4967]: I1008 00:18:21.731662 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/36b9e156-0960-4be2-90b7-8abe8c34d1c7-node-pullsecrets\") pod \"36b9e156-0960-4be2-90b7-8abe8c34d1c7\" (UID: \"36b9e156-0960-4be2-90b7-8abe8c34d1c7\") " Oct 08 00:18:21 crc kubenswrapper[4967]: I1008 00:18:21.731762 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/36b9e156-0960-4be2-90b7-8abe8c34d1c7-buildworkdir\") pod \"36b9e156-0960-4be2-90b7-8abe8c34d1c7\" (UID: \"36b9e156-0960-4be2-90b7-8abe8c34d1c7\") " Oct 08 00:18:21 crc kubenswrapper[4967]: I1008 00:18:21.731814 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-r7snr-pull\" (UniqueName: \"kubernetes.io/secret/36b9e156-0960-4be2-90b7-8abe8c34d1c7-builder-dockercfg-r7snr-pull\") pod \"36b9e156-0960-4be2-90b7-8abe8c34d1c7\" (UID: \"36b9e156-0960-4be2-90b7-8abe8c34d1c7\") " Oct 08 00:18:21 crc kubenswrapper[4967]: I1008 00:18:21.731882 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/36b9e156-0960-4be2-90b7-8abe8c34d1c7-container-storage-root\") pod \"36b9e156-0960-4be2-90b7-8abe8c34d1c7\" (UID: \"36b9e156-0960-4be2-90b7-8abe8c34d1c7\") " Oct 08 00:18:21 crc kubenswrapper[4967]: I1008 00:18:21.731977 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/36b9e156-0960-4be2-90b7-8abe8c34d1c7-build-system-configs\") pod \"36b9e156-0960-4be2-90b7-8abe8c34d1c7\" (UID: \"36b9e156-0960-4be2-90b7-8abe8c34d1c7\") " Oct 08 00:18:21 crc kubenswrapper[4967]: I1008 00:18:21.732058 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/36b9e156-0960-4be2-90b7-8abe8c34d1c7-build-blob-cache\") pod \"36b9e156-0960-4be2-90b7-8abe8c34d1c7\" (UID: \"36b9e156-0960-4be2-90b7-8abe8c34d1c7\") " Oct 08 00:18:21 crc kubenswrapper[4967]: I1008 00:18:21.731809 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/36b9e156-0960-4be2-90b7-8abe8c34d1c7-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "36b9e156-0960-4be2-90b7-8abe8c34d1c7" (UID: "36b9e156-0960-4be2-90b7-8abe8c34d1c7"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 00:18:21 crc kubenswrapper[4967]: I1008 00:18:21.731847 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/36b9e156-0960-4be2-90b7-8abe8c34d1c7-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "36b9e156-0960-4be2-90b7-8abe8c34d1c7" (UID: "36b9e156-0960-4be2-90b7-8abe8c34d1c7"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 00:18:21 crc kubenswrapper[4967]: I1008 00:18:21.732285 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/36b9e156-0960-4be2-90b7-8abe8c34d1c7-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "36b9e156-0960-4be2-90b7-8abe8c34d1c7" (UID: "36b9e156-0960-4be2-90b7-8abe8c34d1c7"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:18:21 crc kubenswrapper[4967]: I1008 00:18:21.732560 4967 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/36b9e156-0960-4be2-90b7-8abe8c34d1c7-buildcachedir\") on node \"crc\" DevicePath \"\"" Oct 08 00:18:21 crc kubenswrapper[4967]: I1008 00:18:21.732606 4967 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/36b9e156-0960-4be2-90b7-8abe8c34d1c7-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Oct 08 00:18:21 crc kubenswrapper[4967]: I1008 00:18:21.732629 4967 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/36b9e156-0960-4be2-90b7-8abe8c34d1c7-buildworkdir\") on node \"crc\" DevicePath \"\"" Oct 08 00:18:21 crc kubenswrapper[4967]: I1008 00:18:21.736393 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36b9e156-0960-4be2-90b7-8abe8c34d1c7-builder-dockercfg-r7snr-pull" (OuterVolumeSpecName: "builder-dockercfg-r7snr-pull") pod "36b9e156-0960-4be2-90b7-8abe8c34d1c7" (UID: "36b9e156-0960-4be2-90b7-8abe8c34d1c7"). InnerVolumeSpecName "builder-dockercfg-r7snr-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:18:21 crc kubenswrapper[4967]: I1008 00:18:21.736830 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/36b9e156-0960-4be2-90b7-8abe8c34d1c7-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "36b9e156-0960-4be2-90b7-8abe8c34d1c7" (UID: "36b9e156-0960-4be2-90b7-8abe8c34d1c7"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:18:21 crc kubenswrapper[4967]: I1008 00:18:21.736843 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/36b9e156-0960-4be2-90b7-8abe8c34d1c7-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "36b9e156-0960-4be2-90b7-8abe8c34d1c7" (UID: "36b9e156-0960-4be2-90b7-8abe8c34d1c7"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:18:21 crc kubenswrapper[4967]: I1008 00:18:21.736896 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/36b9e156-0960-4be2-90b7-8abe8c34d1c7-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "36b9e156-0960-4be2-90b7-8abe8c34d1c7" (UID: "36b9e156-0960-4be2-90b7-8abe8c34d1c7"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:18:21 crc kubenswrapper[4967]: I1008 00:18:21.736931 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/36b9e156-0960-4be2-90b7-8abe8c34d1c7-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "36b9e156-0960-4be2-90b7-8abe8c34d1c7" (UID: "36b9e156-0960-4be2-90b7-8abe8c34d1c7"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:18:21 crc kubenswrapper[4967]: I1008 00:18:21.737120 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/36b9e156-0960-4be2-90b7-8abe8c34d1c7-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "36b9e156-0960-4be2-90b7-8abe8c34d1c7" (UID: "36b9e156-0960-4be2-90b7-8abe8c34d1c7"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:18:21 crc kubenswrapper[4967]: I1008 00:18:21.737382 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/36b9e156-0960-4be2-90b7-8abe8c34d1c7-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "36b9e156-0960-4be2-90b7-8abe8c34d1c7" (UID: "36b9e156-0960-4be2-90b7-8abe8c34d1c7"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:18:21 crc kubenswrapper[4967]: I1008 00:18:21.737604 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36b9e156-0960-4be2-90b7-8abe8c34d1c7-builder-dockercfg-r7snr-push" (OuterVolumeSpecName: "builder-dockercfg-r7snr-push") pod "36b9e156-0960-4be2-90b7-8abe8c34d1c7" (UID: "36b9e156-0960-4be2-90b7-8abe8c34d1c7"). InnerVolumeSpecName "builder-dockercfg-r7snr-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:18:21 crc kubenswrapper[4967]: I1008 00:18:21.738295 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36b9e156-0960-4be2-90b7-8abe8c34d1c7-kube-api-access-hbg9g" (OuterVolumeSpecName: "kube-api-access-hbg9g") pod "36b9e156-0960-4be2-90b7-8abe8c34d1c7" (UID: "36b9e156-0960-4be2-90b7-8abe8c34d1c7"). InnerVolumeSpecName "kube-api-access-hbg9g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:18:21 crc kubenswrapper[4967]: I1008 00:18:21.834716 4967 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/36b9e156-0960-4be2-90b7-8abe8c34d1c7-build-blob-cache\") on node \"crc\" DevicePath \"\"" Oct 08 00:18:21 crc kubenswrapper[4967]: I1008 00:18:21.834793 4967 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-r7snr-push\" (UniqueName: \"kubernetes.io/secret/36b9e156-0960-4be2-90b7-8abe8c34d1c7-builder-dockercfg-r7snr-push\") on node \"crc\" DevicePath \"\"" Oct 08 00:18:21 crc kubenswrapper[4967]: I1008 00:18:21.834826 4967 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/36b9e156-0960-4be2-90b7-8abe8c34d1c7-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 08 00:18:21 crc kubenswrapper[4967]: I1008 00:18:21.834856 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hbg9g\" (UniqueName: \"kubernetes.io/projected/36b9e156-0960-4be2-90b7-8abe8c34d1c7-kube-api-access-hbg9g\") on node \"crc\" DevicePath \"\"" Oct 08 00:18:21 crc kubenswrapper[4967]: I1008 00:18:21.834884 4967 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/36b9e156-0960-4be2-90b7-8abe8c34d1c7-container-storage-run\") on node \"crc\" DevicePath \"\"" Oct 08 00:18:21 crc kubenswrapper[4967]: I1008 00:18:21.834910 4967 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/36b9e156-0960-4be2-90b7-8abe8c34d1c7-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 08 00:18:21 crc kubenswrapper[4967]: I1008 00:18:21.834935 4967 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-r7snr-pull\" (UniqueName: \"kubernetes.io/secret/36b9e156-0960-4be2-90b7-8abe8c34d1c7-builder-dockercfg-r7snr-pull\") on node \"crc\" DevicePath \"\"" Oct 08 00:18:21 crc kubenswrapper[4967]: I1008 00:18:21.834960 4967 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/36b9e156-0960-4be2-90b7-8abe8c34d1c7-container-storage-root\") on node \"crc\" DevicePath \"\"" Oct 08 00:18:21 crc kubenswrapper[4967]: I1008 00:18:21.834978 4967 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/36b9e156-0960-4be2-90b7-8abe8c34d1c7-build-system-configs\") on node \"crc\" DevicePath \"\"" Oct 08 00:18:22 crc kubenswrapper[4967]: I1008 00:18:22.532157 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-1-build_36b9e156-0960-4be2-90b7-8abe8c34d1c7/docker-build/0.log" Oct 08 00:18:22 crc kubenswrapper[4967]: I1008 00:18:22.532838 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-1-build" Oct 08 00:18:22 crc kubenswrapper[4967]: I1008 00:18:22.532832 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-1-build" event={"ID":"36b9e156-0960-4be2-90b7-8abe8c34d1c7","Type":"ContainerDied","Data":"0b69261a4c9913ce438892715d5fb83e32006cb986365f731de58c1e1bbc78d4"} Oct 08 00:18:22 crc kubenswrapper[4967]: I1008 00:18:22.533073 4967 scope.go:117] "RemoveContainer" containerID="59b34e9e905cad4da9574fe662b46abeed4961716c6db3bf78dceab52c3b0d9c" Oct 08 00:18:22 crc kubenswrapper[4967]: I1008 00:18:22.535086 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-2-build" event={"ID":"76078656-a41c-4e34-ad5d-ca11ebaa3ff3","Type":"ContainerStarted","Data":"64229a976ed1fb00da7f6d2d09a5587e5638fa799a1b2c40a6e3bc0a46945225"} Oct 08 00:18:22 crc kubenswrapper[4967]: I1008 00:18:22.565842 4967 scope.go:117] "RemoveContainer" containerID="dd5311ce8b169d403504f774c3492781abc4472d38e9278c261868c44077be83" Oct 08 00:18:22 crc kubenswrapper[4967]: I1008 00:18:22.609231 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/service-telemetry-operator-1-build"] Oct 08 00:18:22 crc kubenswrapper[4967]: I1008 00:18:22.615140 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["service-telemetry/service-telemetry-operator-1-build"] Oct 08 00:18:24 crc kubenswrapper[4967]: I1008 00:18:24.126951 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="36b9e156-0960-4be2-90b7-8abe8c34d1c7" path="/var/lib/kubelet/pods/36b9e156-0960-4be2-90b7-8abe8c34d1c7/volumes" Oct 08 00:18:30 crc kubenswrapper[4967]: I1008 00:18:30.601290 4967 generic.go:334] "Generic (PLEG): container finished" podID="76078656-a41c-4e34-ad5d-ca11ebaa3ff3" containerID="64229a976ed1fb00da7f6d2d09a5587e5638fa799a1b2c40a6e3bc0a46945225" exitCode=0 Oct 08 00:18:30 crc kubenswrapper[4967]: I1008 00:18:30.601416 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-2-build" event={"ID":"76078656-a41c-4e34-ad5d-ca11ebaa3ff3","Type":"ContainerDied","Data":"64229a976ed1fb00da7f6d2d09a5587e5638fa799a1b2c40a6e3bc0a46945225"} Oct 08 00:18:31 crc kubenswrapper[4967]: I1008 00:18:31.609402 4967 generic.go:334] "Generic (PLEG): container finished" podID="76078656-a41c-4e34-ad5d-ca11ebaa3ff3" containerID="60734f743cbc8c50525252164ebe122be9aae8337ab3c7d55d758938fa7113b2" exitCode=0 Oct 08 00:18:31 crc kubenswrapper[4967]: I1008 00:18:31.609491 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-2-build" event={"ID":"76078656-a41c-4e34-ad5d-ca11ebaa3ff3","Type":"ContainerDied","Data":"60734f743cbc8c50525252164ebe122be9aae8337ab3c7d55d758938fa7113b2"} Oct 08 00:18:31 crc kubenswrapper[4967]: I1008 00:18:31.656580 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-2-build_76078656-a41c-4e34-ad5d-ca11ebaa3ff3/manage-dockerfile/0.log" Oct 08 00:18:32 crc kubenswrapper[4967]: I1008 00:18:32.619420 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-2-build" event={"ID":"76078656-a41c-4e34-ad5d-ca11ebaa3ff3","Type":"ContainerStarted","Data":"0a43c71e4932386d7fed98c8560524c1d4b0a845f0edd0e8b48a515f27036226"} Oct 08 00:18:32 crc kubenswrapper[4967]: I1008 00:18:32.658006 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/service-telemetry-operator-2-build" podStartSLOduration=17.657983298 podStartE2EDuration="17.657983298s" podCreationTimestamp="2025-10-08 00:18:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:18:32.656191185 +0000 UTC m=+711.306725864" watchObservedRunningTime="2025-10-08 00:18:32.657983298 +0000 UTC m=+711.308517967" Oct 08 00:18:44 crc kubenswrapper[4967]: I1008 00:18:44.255615 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-vkf46"] Oct 08 00:18:44 crc kubenswrapper[4967]: I1008 00:18:44.256295 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-vkf46" podUID="95dc52d1-8ca7-45ca-83c4-0eedafd40085" containerName="controller-manager" containerID="cri-o://4106e7ce0e367c0ecf5bf205d01584d073e67ff15fcc1437538cbd2d462cc931" gracePeriod=30 Oct 08 00:18:44 crc kubenswrapper[4967]: I1008 00:18:44.353647 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-v4kxw"] Oct 08 00:18:44 crc kubenswrapper[4967]: I1008 00:18:44.353870 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v4kxw" podUID="2950f912-e1e2-4640-b946-fd9ac7f2a8d3" containerName="route-controller-manager" containerID="cri-o://4388470ff7f5a8b15c206786fa0b888701511b6432b249e45dcbe25207ebf21c" gracePeriod=30 Oct 08 00:18:44 crc kubenswrapper[4967]: I1008 00:18:44.697402 4967 generic.go:334] "Generic (PLEG): container finished" podID="95dc52d1-8ca7-45ca-83c4-0eedafd40085" containerID="4106e7ce0e367c0ecf5bf205d01584d073e67ff15fcc1437538cbd2d462cc931" exitCode=0 Oct 08 00:18:44 crc kubenswrapper[4967]: I1008 00:18:44.697481 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-vkf46" event={"ID":"95dc52d1-8ca7-45ca-83c4-0eedafd40085","Type":"ContainerDied","Data":"4106e7ce0e367c0ecf5bf205d01584d073e67ff15fcc1437538cbd2d462cc931"} Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.162900 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-vkf46" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.245622 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v4kxw" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.266484 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2950f912-e1e2-4640-b946-fd9ac7f2a8d3-config\") pod \"2950f912-e1e2-4640-b946-fd9ac7f2a8d3\" (UID: \"2950f912-e1e2-4640-b946-fd9ac7f2a8d3\") " Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.266548 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/95dc52d1-8ca7-45ca-83c4-0eedafd40085-config\") pod \"95dc52d1-8ca7-45ca-83c4-0eedafd40085\" (UID: \"95dc52d1-8ca7-45ca-83c4-0eedafd40085\") " Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.266574 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fsmpw\" (UniqueName: \"kubernetes.io/projected/95dc52d1-8ca7-45ca-83c4-0eedafd40085-kube-api-access-fsmpw\") pod \"95dc52d1-8ca7-45ca-83c4-0eedafd40085\" (UID: \"95dc52d1-8ca7-45ca-83c4-0eedafd40085\") " Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.266615 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bjxsg\" (UniqueName: \"kubernetes.io/projected/2950f912-e1e2-4640-b946-fd9ac7f2a8d3-kube-api-access-bjxsg\") pod \"2950f912-e1e2-4640-b946-fd9ac7f2a8d3\" (UID: \"2950f912-e1e2-4640-b946-fd9ac7f2a8d3\") " Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.266689 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2950f912-e1e2-4640-b946-fd9ac7f2a8d3-client-ca\") pod \"2950f912-e1e2-4640-b946-fd9ac7f2a8d3\" (UID: \"2950f912-e1e2-4640-b946-fd9ac7f2a8d3\") " Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.266747 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/95dc52d1-8ca7-45ca-83c4-0eedafd40085-serving-cert\") pod \"95dc52d1-8ca7-45ca-83c4-0eedafd40085\" (UID: \"95dc52d1-8ca7-45ca-83c4-0eedafd40085\") " Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.266778 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2950f912-e1e2-4640-b946-fd9ac7f2a8d3-serving-cert\") pod \"2950f912-e1e2-4640-b946-fd9ac7f2a8d3\" (UID: \"2950f912-e1e2-4640-b946-fd9ac7f2a8d3\") " Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.266818 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/95dc52d1-8ca7-45ca-83c4-0eedafd40085-client-ca\") pod \"95dc52d1-8ca7-45ca-83c4-0eedafd40085\" (UID: \"95dc52d1-8ca7-45ca-83c4-0eedafd40085\") " Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.266839 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/95dc52d1-8ca7-45ca-83c4-0eedafd40085-proxy-ca-bundles\") pod \"95dc52d1-8ca7-45ca-83c4-0eedafd40085\" (UID: \"95dc52d1-8ca7-45ca-83c4-0eedafd40085\") " Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.267498 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2950f912-e1e2-4640-b946-fd9ac7f2a8d3-config" (OuterVolumeSpecName: "config") pod "2950f912-e1e2-4640-b946-fd9ac7f2a8d3" (UID: "2950f912-e1e2-4640-b946-fd9ac7f2a8d3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.267628 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/95dc52d1-8ca7-45ca-83c4-0eedafd40085-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "95dc52d1-8ca7-45ca-83c4-0eedafd40085" (UID: "95dc52d1-8ca7-45ca-83c4-0eedafd40085"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.267738 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/95dc52d1-8ca7-45ca-83c4-0eedafd40085-config" (OuterVolumeSpecName: "config") pod "95dc52d1-8ca7-45ca-83c4-0eedafd40085" (UID: "95dc52d1-8ca7-45ca-83c4-0eedafd40085"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.268078 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2950f912-e1e2-4640-b946-fd9ac7f2a8d3-client-ca" (OuterVolumeSpecName: "client-ca") pod "2950f912-e1e2-4640-b946-fd9ac7f2a8d3" (UID: "2950f912-e1e2-4640-b946-fd9ac7f2a8d3"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.268337 4967 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2950f912-e1e2-4640-b946-fd9ac7f2a8d3-config\") on node \"crc\" DevicePath \"\"" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.268363 4967 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/95dc52d1-8ca7-45ca-83c4-0eedafd40085-config\") on node \"crc\" DevicePath \"\"" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.268377 4967 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2950f912-e1e2-4640-b946-fd9ac7f2a8d3-client-ca\") on node \"crc\" DevicePath \"\"" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.268416 4967 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/95dc52d1-8ca7-45ca-83c4-0eedafd40085-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.268968 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/95dc52d1-8ca7-45ca-83c4-0eedafd40085-client-ca" (OuterVolumeSpecName: "client-ca") pod "95dc52d1-8ca7-45ca-83c4-0eedafd40085" (UID: "95dc52d1-8ca7-45ca-83c4-0eedafd40085"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.273772 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2950f912-e1e2-4640-b946-fd9ac7f2a8d3-kube-api-access-bjxsg" (OuterVolumeSpecName: "kube-api-access-bjxsg") pod "2950f912-e1e2-4640-b946-fd9ac7f2a8d3" (UID: "2950f912-e1e2-4640-b946-fd9ac7f2a8d3"). InnerVolumeSpecName "kube-api-access-bjxsg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.273865 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95dc52d1-8ca7-45ca-83c4-0eedafd40085-kube-api-access-fsmpw" (OuterVolumeSpecName: "kube-api-access-fsmpw") pod "95dc52d1-8ca7-45ca-83c4-0eedafd40085" (UID: "95dc52d1-8ca7-45ca-83c4-0eedafd40085"). InnerVolumeSpecName "kube-api-access-fsmpw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.274427 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2950f912-e1e2-4640-b946-fd9ac7f2a8d3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "2950f912-e1e2-4640-b946-fd9ac7f2a8d3" (UID: "2950f912-e1e2-4640-b946-fd9ac7f2a8d3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.279156 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95dc52d1-8ca7-45ca-83c4-0eedafd40085-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "95dc52d1-8ca7-45ca-83c4-0eedafd40085" (UID: "95dc52d1-8ca7-45ca-83c4-0eedafd40085"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.369011 4967 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/95dc52d1-8ca7-45ca-83c4-0eedafd40085-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.369213 4967 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2950f912-e1e2-4640-b946-fd9ac7f2a8d3-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.369222 4967 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/95dc52d1-8ca7-45ca-83c4-0eedafd40085-client-ca\") on node \"crc\" DevicePath \"\"" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.369231 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fsmpw\" (UniqueName: \"kubernetes.io/projected/95dc52d1-8ca7-45ca-83c4-0eedafd40085-kube-api-access-fsmpw\") on node \"crc\" DevicePath \"\"" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.369239 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bjxsg\" (UniqueName: \"kubernetes.io/projected/2950f912-e1e2-4640-b946-fd9ac7f2a8d3-kube-api-access-bjxsg\") on node \"crc\" DevicePath \"\"" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.422982 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-8586d5c566-n9862"] Oct 08 00:18:45 crc kubenswrapper[4967]: E1008 00:18:45.423304 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36b9e156-0960-4be2-90b7-8abe8c34d1c7" containerName="docker-build" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.423322 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="36b9e156-0960-4be2-90b7-8abe8c34d1c7" containerName="docker-build" Oct 08 00:18:45 crc kubenswrapper[4967]: E1008 00:18:45.423337 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36b9e156-0960-4be2-90b7-8abe8c34d1c7" containerName="manage-dockerfile" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.423346 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="36b9e156-0960-4be2-90b7-8abe8c34d1c7" containerName="manage-dockerfile" Oct 08 00:18:45 crc kubenswrapper[4967]: E1008 00:18:45.423368 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2950f912-e1e2-4640-b946-fd9ac7f2a8d3" containerName="route-controller-manager" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.423377 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="2950f912-e1e2-4640-b946-fd9ac7f2a8d3" containerName="route-controller-manager" Oct 08 00:18:45 crc kubenswrapper[4967]: E1008 00:18:45.423390 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95dc52d1-8ca7-45ca-83c4-0eedafd40085" containerName="controller-manager" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.423397 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="95dc52d1-8ca7-45ca-83c4-0eedafd40085" containerName="controller-manager" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.423517 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="2950f912-e1e2-4640-b946-fd9ac7f2a8d3" containerName="route-controller-manager" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.423531 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="36b9e156-0960-4be2-90b7-8abe8c34d1c7" containerName="docker-build" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.423540 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="95dc52d1-8ca7-45ca-83c4-0eedafd40085" containerName="controller-manager" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.423991 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-8586d5c566-n9862" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.427186 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-76c56976c5-rh49w"] Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.428061 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-76c56976c5-rh49w" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.437223 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-8586d5c566-n9862"] Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.441026 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-76c56976c5-rh49w"] Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.470874 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8950261e-9b3f-443e-ab30-692f2604185a-serving-cert\") pod \"controller-manager-8586d5c566-n9862\" (UID: \"8950261e-9b3f-443e-ab30-692f2604185a\") " pod="openshift-controller-manager/controller-manager-8586d5c566-n9862" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.470945 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d9e5f3be-2a1f-4890-acf7-6a9dae8f1f5a-config\") pod \"route-controller-manager-76c56976c5-rh49w\" (UID: \"d9e5f3be-2a1f-4890-acf7-6a9dae8f1f5a\") " pod="openshift-route-controller-manager/route-controller-manager-76c56976c5-rh49w" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.471029 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d9e5f3be-2a1f-4890-acf7-6a9dae8f1f5a-client-ca\") pod \"route-controller-manager-76c56976c5-rh49w\" (UID: \"d9e5f3be-2a1f-4890-acf7-6a9dae8f1f5a\") " pod="openshift-route-controller-manager/route-controller-manager-76c56976c5-rh49w" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.471079 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mcqrp\" (UniqueName: \"kubernetes.io/projected/d9e5f3be-2a1f-4890-acf7-6a9dae8f1f5a-kube-api-access-mcqrp\") pod \"route-controller-manager-76c56976c5-rh49w\" (UID: \"d9e5f3be-2a1f-4890-acf7-6a9dae8f1f5a\") " pod="openshift-route-controller-manager/route-controller-manager-76c56976c5-rh49w" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.471118 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8950261e-9b3f-443e-ab30-692f2604185a-client-ca\") pod \"controller-manager-8586d5c566-n9862\" (UID: \"8950261e-9b3f-443e-ab30-692f2604185a\") " pod="openshift-controller-manager/controller-manager-8586d5c566-n9862" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.471149 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8950261e-9b3f-443e-ab30-692f2604185a-config\") pod \"controller-manager-8586d5c566-n9862\" (UID: \"8950261e-9b3f-443e-ab30-692f2604185a\") " pod="openshift-controller-manager/controller-manager-8586d5c566-n9862" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.471257 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8950261e-9b3f-443e-ab30-692f2604185a-proxy-ca-bundles\") pod \"controller-manager-8586d5c566-n9862\" (UID: \"8950261e-9b3f-443e-ab30-692f2604185a\") " pod="openshift-controller-manager/controller-manager-8586d5c566-n9862" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.471303 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d9e5f3be-2a1f-4890-acf7-6a9dae8f1f5a-serving-cert\") pod \"route-controller-manager-76c56976c5-rh49w\" (UID: \"d9e5f3be-2a1f-4890-acf7-6a9dae8f1f5a\") " pod="openshift-route-controller-manager/route-controller-manager-76c56976c5-rh49w" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.471429 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9dh68\" (UniqueName: \"kubernetes.io/projected/8950261e-9b3f-443e-ab30-692f2604185a-kube-api-access-9dh68\") pod \"controller-manager-8586d5c566-n9862\" (UID: \"8950261e-9b3f-443e-ab30-692f2604185a\") " pod="openshift-controller-manager/controller-manager-8586d5c566-n9862" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.486297 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-8586d5c566-n9862"] Oct 08 00:18:45 crc kubenswrapper[4967]: E1008 00:18:45.486755 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[client-ca config kube-api-access-9dh68 proxy-ca-bundles serving-cert], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openshift-controller-manager/controller-manager-8586d5c566-n9862" podUID="8950261e-9b3f-443e-ab30-692f2604185a" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.499963 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-76c56976c5-rh49w"] Oct 08 00:18:45 crc kubenswrapper[4967]: E1008 00:18:45.500454 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[client-ca config kube-api-access-mcqrp serving-cert], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openshift-route-controller-manager/route-controller-manager-76c56976c5-rh49w" podUID="d9e5f3be-2a1f-4890-acf7-6a9dae8f1f5a" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.572723 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d9e5f3be-2a1f-4890-acf7-6a9dae8f1f5a-config\") pod \"route-controller-manager-76c56976c5-rh49w\" (UID: \"d9e5f3be-2a1f-4890-acf7-6a9dae8f1f5a\") " pod="openshift-route-controller-manager/route-controller-manager-76c56976c5-rh49w" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.572769 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d9e5f3be-2a1f-4890-acf7-6a9dae8f1f5a-client-ca\") pod \"route-controller-manager-76c56976c5-rh49w\" (UID: \"d9e5f3be-2a1f-4890-acf7-6a9dae8f1f5a\") " pod="openshift-route-controller-manager/route-controller-manager-76c56976c5-rh49w" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.572793 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mcqrp\" (UniqueName: \"kubernetes.io/projected/d9e5f3be-2a1f-4890-acf7-6a9dae8f1f5a-kube-api-access-mcqrp\") pod \"route-controller-manager-76c56976c5-rh49w\" (UID: \"d9e5f3be-2a1f-4890-acf7-6a9dae8f1f5a\") " pod="openshift-route-controller-manager/route-controller-manager-76c56976c5-rh49w" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.572812 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8950261e-9b3f-443e-ab30-692f2604185a-client-ca\") pod \"controller-manager-8586d5c566-n9862\" (UID: \"8950261e-9b3f-443e-ab30-692f2604185a\") " pod="openshift-controller-manager/controller-manager-8586d5c566-n9862" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.572831 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8950261e-9b3f-443e-ab30-692f2604185a-config\") pod \"controller-manager-8586d5c566-n9862\" (UID: \"8950261e-9b3f-443e-ab30-692f2604185a\") " pod="openshift-controller-manager/controller-manager-8586d5c566-n9862" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.572862 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8950261e-9b3f-443e-ab30-692f2604185a-proxy-ca-bundles\") pod \"controller-manager-8586d5c566-n9862\" (UID: \"8950261e-9b3f-443e-ab30-692f2604185a\") " pod="openshift-controller-manager/controller-manager-8586d5c566-n9862" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.572882 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d9e5f3be-2a1f-4890-acf7-6a9dae8f1f5a-serving-cert\") pod \"route-controller-manager-76c56976c5-rh49w\" (UID: \"d9e5f3be-2a1f-4890-acf7-6a9dae8f1f5a\") " pod="openshift-route-controller-manager/route-controller-manager-76c56976c5-rh49w" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.572933 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9dh68\" (UniqueName: \"kubernetes.io/projected/8950261e-9b3f-443e-ab30-692f2604185a-kube-api-access-9dh68\") pod \"controller-manager-8586d5c566-n9862\" (UID: \"8950261e-9b3f-443e-ab30-692f2604185a\") " pod="openshift-controller-manager/controller-manager-8586d5c566-n9862" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.572961 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8950261e-9b3f-443e-ab30-692f2604185a-serving-cert\") pod \"controller-manager-8586d5c566-n9862\" (UID: \"8950261e-9b3f-443e-ab30-692f2604185a\") " pod="openshift-controller-manager/controller-manager-8586d5c566-n9862" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.574036 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d9e5f3be-2a1f-4890-acf7-6a9dae8f1f5a-client-ca\") pod \"route-controller-manager-76c56976c5-rh49w\" (UID: \"d9e5f3be-2a1f-4890-acf7-6a9dae8f1f5a\") " pod="openshift-route-controller-manager/route-controller-manager-76c56976c5-rh49w" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.574041 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8950261e-9b3f-443e-ab30-692f2604185a-client-ca\") pod \"controller-manager-8586d5c566-n9862\" (UID: \"8950261e-9b3f-443e-ab30-692f2604185a\") " pod="openshift-controller-manager/controller-manager-8586d5c566-n9862" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.575196 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d9e5f3be-2a1f-4890-acf7-6a9dae8f1f5a-config\") pod \"route-controller-manager-76c56976c5-rh49w\" (UID: \"d9e5f3be-2a1f-4890-acf7-6a9dae8f1f5a\") " pod="openshift-route-controller-manager/route-controller-manager-76c56976c5-rh49w" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.575777 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8950261e-9b3f-443e-ab30-692f2604185a-proxy-ca-bundles\") pod \"controller-manager-8586d5c566-n9862\" (UID: \"8950261e-9b3f-443e-ab30-692f2604185a\") " pod="openshift-controller-manager/controller-manager-8586d5c566-n9862" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.576329 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8950261e-9b3f-443e-ab30-692f2604185a-config\") pod \"controller-manager-8586d5c566-n9862\" (UID: \"8950261e-9b3f-443e-ab30-692f2604185a\") " pod="openshift-controller-manager/controller-manager-8586d5c566-n9862" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.579607 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8950261e-9b3f-443e-ab30-692f2604185a-serving-cert\") pod \"controller-manager-8586d5c566-n9862\" (UID: \"8950261e-9b3f-443e-ab30-692f2604185a\") " pod="openshift-controller-manager/controller-manager-8586d5c566-n9862" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.579714 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d9e5f3be-2a1f-4890-acf7-6a9dae8f1f5a-serving-cert\") pod \"route-controller-manager-76c56976c5-rh49w\" (UID: \"d9e5f3be-2a1f-4890-acf7-6a9dae8f1f5a\") " pod="openshift-route-controller-manager/route-controller-manager-76c56976c5-rh49w" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.593423 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9dh68\" (UniqueName: \"kubernetes.io/projected/8950261e-9b3f-443e-ab30-692f2604185a-kube-api-access-9dh68\") pod \"controller-manager-8586d5c566-n9862\" (UID: \"8950261e-9b3f-443e-ab30-692f2604185a\") " pod="openshift-controller-manager/controller-manager-8586d5c566-n9862" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.596975 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mcqrp\" (UniqueName: \"kubernetes.io/projected/d9e5f3be-2a1f-4890-acf7-6a9dae8f1f5a-kube-api-access-mcqrp\") pod \"route-controller-manager-76c56976c5-rh49w\" (UID: \"d9e5f3be-2a1f-4890-acf7-6a9dae8f1f5a\") " pod="openshift-route-controller-manager/route-controller-manager-76c56976c5-rh49w" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.704865 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-vkf46" event={"ID":"95dc52d1-8ca7-45ca-83c4-0eedafd40085","Type":"ContainerDied","Data":"17c7e7c8e033ebbfecd21ef1a5396727c0b3abcdb20bfefac115ede13cbfc25f"} Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.704918 4967 scope.go:117] "RemoveContainer" containerID="4106e7ce0e367c0ecf5bf205d01584d073e67ff15fcc1437538cbd2d462cc931" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.705029 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-vkf46" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.718809 4967 generic.go:334] "Generic (PLEG): container finished" podID="2950f912-e1e2-4640-b946-fd9ac7f2a8d3" containerID="4388470ff7f5a8b15c206786fa0b888701511b6432b249e45dcbe25207ebf21c" exitCode=0 Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.718888 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-8586d5c566-n9862" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.719369 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v4kxw" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.723720 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v4kxw" event={"ID":"2950f912-e1e2-4640-b946-fd9ac7f2a8d3","Type":"ContainerDied","Data":"4388470ff7f5a8b15c206786fa0b888701511b6432b249e45dcbe25207ebf21c"} Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.723791 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-v4kxw" event={"ID":"2950f912-e1e2-4640-b946-fd9ac7f2a8d3","Type":"ContainerDied","Data":"5b0455f52fa7244be50a4ba95d0f95417f00483fea24320be312e4c8d08c2a77"} Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.723810 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-76c56976c5-rh49w" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.740106 4967 scope.go:117] "RemoveContainer" containerID="4388470ff7f5a8b15c206786fa0b888701511b6432b249e45dcbe25207ebf21c" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.753063 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-8586d5c566-n9862" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.753159 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-76c56976c5-rh49w" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.768885 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-vkf46"] Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.772920 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-vkf46"] Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.775105 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8950261e-9b3f-443e-ab30-692f2604185a-client-ca\") pod \"8950261e-9b3f-443e-ab30-692f2604185a\" (UID: \"8950261e-9b3f-443e-ab30-692f2604185a\") " Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.775153 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9dh68\" (UniqueName: \"kubernetes.io/projected/8950261e-9b3f-443e-ab30-692f2604185a-kube-api-access-9dh68\") pod \"8950261e-9b3f-443e-ab30-692f2604185a\" (UID: \"8950261e-9b3f-443e-ab30-692f2604185a\") " Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.775189 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d9e5f3be-2a1f-4890-acf7-6a9dae8f1f5a-client-ca\") pod \"d9e5f3be-2a1f-4890-acf7-6a9dae8f1f5a\" (UID: \"d9e5f3be-2a1f-4890-acf7-6a9dae8f1f5a\") " Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.775233 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d9e5f3be-2a1f-4890-acf7-6a9dae8f1f5a-config\") pod \"d9e5f3be-2a1f-4890-acf7-6a9dae8f1f5a\" (UID: \"d9e5f3be-2a1f-4890-acf7-6a9dae8f1f5a\") " Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.775290 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8950261e-9b3f-443e-ab30-692f2604185a-serving-cert\") pod \"8950261e-9b3f-443e-ab30-692f2604185a\" (UID: \"8950261e-9b3f-443e-ab30-692f2604185a\") " Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.775375 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d9e5f3be-2a1f-4890-acf7-6a9dae8f1f5a-serving-cert\") pod \"d9e5f3be-2a1f-4890-acf7-6a9dae8f1f5a\" (UID: \"d9e5f3be-2a1f-4890-acf7-6a9dae8f1f5a\") " Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.775402 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8950261e-9b3f-443e-ab30-692f2604185a-config\") pod \"8950261e-9b3f-443e-ab30-692f2604185a\" (UID: \"8950261e-9b3f-443e-ab30-692f2604185a\") " Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.775438 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8950261e-9b3f-443e-ab30-692f2604185a-proxy-ca-bundles\") pod \"8950261e-9b3f-443e-ab30-692f2604185a\" (UID: \"8950261e-9b3f-443e-ab30-692f2604185a\") " Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.775455 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mcqrp\" (UniqueName: \"kubernetes.io/projected/d9e5f3be-2a1f-4890-acf7-6a9dae8f1f5a-kube-api-access-mcqrp\") pod \"d9e5f3be-2a1f-4890-acf7-6a9dae8f1f5a\" (UID: \"d9e5f3be-2a1f-4890-acf7-6a9dae8f1f5a\") " Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.776959 4967 scope.go:117] "RemoveContainer" containerID="4388470ff7f5a8b15c206786fa0b888701511b6432b249e45dcbe25207ebf21c" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.777518 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d9e5f3be-2a1f-4890-acf7-6a9dae8f1f5a-config" (OuterVolumeSpecName: "config") pod "d9e5f3be-2a1f-4890-acf7-6a9dae8f1f5a" (UID: "d9e5f3be-2a1f-4890-acf7-6a9dae8f1f5a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.777987 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8950261e-9b3f-443e-ab30-692f2604185a-client-ca" (OuterVolumeSpecName: "client-ca") pod "8950261e-9b3f-443e-ab30-692f2604185a" (UID: "8950261e-9b3f-443e-ab30-692f2604185a"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.778190 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8950261e-9b3f-443e-ab30-692f2604185a-config" (OuterVolumeSpecName: "config") pod "8950261e-9b3f-443e-ab30-692f2604185a" (UID: "8950261e-9b3f-443e-ab30-692f2604185a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.780060 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8950261e-9b3f-443e-ab30-692f2604185a-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "8950261e-9b3f-443e-ab30-692f2604185a" (UID: "8950261e-9b3f-443e-ab30-692f2604185a"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:18:45 crc kubenswrapper[4967]: E1008 00:18:45.780199 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4388470ff7f5a8b15c206786fa0b888701511b6432b249e45dcbe25207ebf21c\": container with ID starting with 4388470ff7f5a8b15c206786fa0b888701511b6432b249e45dcbe25207ebf21c not found: ID does not exist" containerID="4388470ff7f5a8b15c206786fa0b888701511b6432b249e45dcbe25207ebf21c" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.780277 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4388470ff7f5a8b15c206786fa0b888701511b6432b249e45dcbe25207ebf21c"} err="failed to get container status \"4388470ff7f5a8b15c206786fa0b888701511b6432b249e45dcbe25207ebf21c\": rpc error: code = NotFound desc = could not find container \"4388470ff7f5a8b15c206786fa0b888701511b6432b249e45dcbe25207ebf21c\": container with ID starting with 4388470ff7f5a8b15c206786fa0b888701511b6432b249e45dcbe25207ebf21c not found: ID does not exist" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.780200 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d9e5f3be-2a1f-4890-acf7-6a9dae8f1f5a-client-ca" (OuterVolumeSpecName: "client-ca") pod "d9e5f3be-2a1f-4890-acf7-6a9dae8f1f5a" (UID: "d9e5f3be-2a1f-4890-acf7-6a9dae8f1f5a"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.806847 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9e5f3be-2a1f-4890-acf7-6a9dae8f1f5a-kube-api-access-mcqrp" (OuterVolumeSpecName: "kube-api-access-mcqrp") pod "d9e5f3be-2a1f-4890-acf7-6a9dae8f1f5a" (UID: "d9e5f3be-2a1f-4890-acf7-6a9dae8f1f5a"). InnerVolumeSpecName "kube-api-access-mcqrp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.810879 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8950261e-9b3f-443e-ab30-692f2604185a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8950261e-9b3f-443e-ab30-692f2604185a" (UID: "8950261e-9b3f-443e-ab30-692f2604185a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.811323 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8950261e-9b3f-443e-ab30-692f2604185a-kube-api-access-9dh68" (OuterVolumeSpecName: "kube-api-access-9dh68") pod "8950261e-9b3f-443e-ab30-692f2604185a" (UID: "8950261e-9b3f-443e-ab30-692f2604185a"). InnerVolumeSpecName "kube-api-access-9dh68". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.813888 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9e5f3be-2a1f-4890-acf7-6a9dae8f1f5a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "d9e5f3be-2a1f-4890-acf7-6a9dae8f1f5a" (UID: "d9e5f3be-2a1f-4890-acf7-6a9dae8f1f5a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.830727 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-v4kxw"] Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.837334 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-v4kxw"] Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.877214 4967 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d9e5f3be-2a1f-4890-acf7-6a9dae8f1f5a-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.877245 4967 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8950261e-9b3f-443e-ab30-692f2604185a-config\") on node \"crc\" DevicePath \"\"" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.877254 4967 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8950261e-9b3f-443e-ab30-692f2604185a-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.877265 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mcqrp\" (UniqueName: \"kubernetes.io/projected/d9e5f3be-2a1f-4890-acf7-6a9dae8f1f5a-kube-api-access-mcqrp\") on node \"crc\" DevicePath \"\"" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.877274 4967 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8950261e-9b3f-443e-ab30-692f2604185a-client-ca\") on node \"crc\" DevicePath \"\"" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.877283 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9dh68\" (UniqueName: \"kubernetes.io/projected/8950261e-9b3f-443e-ab30-692f2604185a-kube-api-access-9dh68\") on node \"crc\" DevicePath \"\"" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.877292 4967 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d9e5f3be-2a1f-4890-acf7-6a9dae8f1f5a-client-ca\") on node \"crc\" DevicePath \"\"" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.877301 4967 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d9e5f3be-2a1f-4890-acf7-6a9dae8f1f5a-config\") on node \"crc\" DevicePath \"\"" Oct 08 00:18:45 crc kubenswrapper[4967]: I1008 00:18:45.877310 4967 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8950261e-9b3f-443e-ab30-692f2604185a-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 00:18:46 crc kubenswrapper[4967]: I1008 00:18:46.128705 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2950f912-e1e2-4640-b946-fd9ac7f2a8d3" path="/var/lib/kubelet/pods/2950f912-e1e2-4640-b946-fd9ac7f2a8d3/volumes" Oct 08 00:18:46 crc kubenswrapper[4967]: I1008 00:18:46.129838 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95dc52d1-8ca7-45ca-83c4-0eedafd40085" path="/var/lib/kubelet/pods/95dc52d1-8ca7-45ca-83c4-0eedafd40085/volumes" Oct 08 00:18:46 crc kubenswrapper[4967]: I1008 00:18:46.726052 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-76c56976c5-rh49w" Oct 08 00:18:46 crc kubenswrapper[4967]: I1008 00:18:46.726105 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-8586d5c566-n9862" Oct 08 00:18:46 crc kubenswrapper[4967]: I1008 00:18:46.775160 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-5cf999dc48-9csf2"] Oct 08 00:18:46 crc kubenswrapper[4967]: I1008 00:18:46.776278 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5cf999dc48-9csf2" Oct 08 00:18:46 crc kubenswrapper[4967]: I1008 00:18:46.779784 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 08 00:18:46 crc kubenswrapper[4967]: I1008 00:18:46.780067 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 08 00:18:46 crc kubenswrapper[4967]: I1008 00:18:46.780491 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 08 00:18:46 crc kubenswrapper[4967]: I1008 00:18:46.780773 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 08 00:18:46 crc kubenswrapper[4967]: I1008 00:18:46.781984 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-8586d5c566-n9862"] Oct 08 00:18:46 crc kubenswrapper[4967]: I1008 00:18:46.783858 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 08 00:18:46 crc kubenswrapper[4967]: I1008 00:18:46.784410 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 08 00:18:46 crc kubenswrapper[4967]: I1008 00:18:46.789031 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/18a6aea8-a3a2-4e08-848b-b33ead229eea-proxy-ca-bundles\") pod \"controller-manager-5cf999dc48-9csf2\" (UID: \"18a6aea8-a3a2-4e08-848b-b33ead229eea\") " pod="openshift-controller-manager/controller-manager-5cf999dc48-9csf2" Oct 08 00:18:46 crc kubenswrapper[4967]: I1008 00:18:46.789273 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lw9xq\" (UniqueName: \"kubernetes.io/projected/18a6aea8-a3a2-4e08-848b-b33ead229eea-kube-api-access-lw9xq\") pod \"controller-manager-5cf999dc48-9csf2\" (UID: \"18a6aea8-a3a2-4e08-848b-b33ead229eea\") " pod="openshift-controller-manager/controller-manager-5cf999dc48-9csf2" Oct 08 00:18:46 crc kubenswrapper[4967]: I1008 00:18:46.789397 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/18a6aea8-a3a2-4e08-848b-b33ead229eea-serving-cert\") pod \"controller-manager-5cf999dc48-9csf2\" (UID: \"18a6aea8-a3a2-4e08-848b-b33ead229eea\") " pod="openshift-controller-manager/controller-manager-5cf999dc48-9csf2" Oct 08 00:18:46 crc kubenswrapper[4967]: I1008 00:18:46.789521 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/18a6aea8-a3a2-4e08-848b-b33ead229eea-client-ca\") pod \"controller-manager-5cf999dc48-9csf2\" (UID: \"18a6aea8-a3a2-4e08-848b-b33ead229eea\") " pod="openshift-controller-manager/controller-manager-5cf999dc48-9csf2" Oct 08 00:18:46 crc kubenswrapper[4967]: I1008 00:18:46.789629 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18a6aea8-a3a2-4e08-848b-b33ead229eea-config\") pod \"controller-manager-5cf999dc48-9csf2\" (UID: \"18a6aea8-a3a2-4e08-848b-b33ead229eea\") " pod="openshift-controller-manager/controller-manager-5cf999dc48-9csf2" Oct 08 00:18:46 crc kubenswrapper[4967]: I1008 00:18:46.801203 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 08 00:18:46 crc kubenswrapper[4967]: I1008 00:18:46.817229 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-8586d5c566-n9862"] Oct 08 00:18:46 crc kubenswrapper[4967]: I1008 00:18:46.835392 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5cf999dc48-9csf2"] Oct 08 00:18:46 crc kubenswrapper[4967]: I1008 00:18:46.848566 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-76c56976c5-rh49w"] Oct 08 00:18:46 crc kubenswrapper[4967]: I1008 00:18:46.857844 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-76c56976c5-rh49w"] Oct 08 00:18:46 crc kubenswrapper[4967]: I1008 00:18:46.890637 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/18a6aea8-a3a2-4e08-848b-b33ead229eea-client-ca\") pod \"controller-manager-5cf999dc48-9csf2\" (UID: \"18a6aea8-a3a2-4e08-848b-b33ead229eea\") " pod="openshift-controller-manager/controller-manager-5cf999dc48-9csf2" Oct 08 00:18:46 crc kubenswrapper[4967]: I1008 00:18:46.890785 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18a6aea8-a3a2-4e08-848b-b33ead229eea-config\") pod \"controller-manager-5cf999dc48-9csf2\" (UID: \"18a6aea8-a3a2-4e08-848b-b33ead229eea\") " pod="openshift-controller-manager/controller-manager-5cf999dc48-9csf2" Oct 08 00:18:46 crc kubenswrapper[4967]: I1008 00:18:46.890912 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/18a6aea8-a3a2-4e08-848b-b33ead229eea-proxy-ca-bundles\") pod \"controller-manager-5cf999dc48-9csf2\" (UID: \"18a6aea8-a3a2-4e08-848b-b33ead229eea\") " pod="openshift-controller-manager/controller-manager-5cf999dc48-9csf2" Oct 08 00:18:46 crc kubenswrapper[4967]: I1008 00:18:46.890996 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lw9xq\" (UniqueName: \"kubernetes.io/projected/18a6aea8-a3a2-4e08-848b-b33ead229eea-kube-api-access-lw9xq\") pod \"controller-manager-5cf999dc48-9csf2\" (UID: \"18a6aea8-a3a2-4e08-848b-b33ead229eea\") " pod="openshift-controller-manager/controller-manager-5cf999dc48-9csf2" Oct 08 00:18:46 crc kubenswrapper[4967]: I1008 00:18:46.891082 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/18a6aea8-a3a2-4e08-848b-b33ead229eea-serving-cert\") pod \"controller-manager-5cf999dc48-9csf2\" (UID: \"18a6aea8-a3a2-4e08-848b-b33ead229eea\") " pod="openshift-controller-manager/controller-manager-5cf999dc48-9csf2" Oct 08 00:18:46 crc kubenswrapper[4967]: I1008 00:18:46.892071 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/18a6aea8-a3a2-4e08-848b-b33ead229eea-client-ca\") pod \"controller-manager-5cf999dc48-9csf2\" (UID: \"18a6aea8-a3a2-4e08-848b-b33ead229eea\") " pod="openshift-controller-manager/controller-manager-5cf999dc48-9csf2" Oct 08 00:18:46 crc kubenswrapper[4967]: I1008 00:18:46.893039 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/18a6aea8-a3a2-4e08-848b-b33ead229eea-proxy-ca-bundles\") pod \"controller-manager-5cf999dc48-9csf2\" (UID: \"18a6aea8-a3a2-4e08-848b-b33ead229eea\") " pod="openshift-controller-manager/controller-manager-5cf999dc48-9csf2" Oct 08 00:18:46 crc kubenswrapper[4967]: I1008 00:18:46.894391 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18a6aea8-a3a2-4e08-848b-b33ead229eea-config\") pod \"controller-manager-5cf999dc48-9csf2\" (UID: \"18a6aea8-a3a2-4e08-848b-b33ead229eea\") " pod="openshift-controller-manager/controller-manager-5cf999dc48-9csf2" Oct 08 00:18:46 crc kubenswrapper[4967]: I1008 00:18:46.896631 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/18a6aea8-a3a2-4e08-848b-b33ead229eea-serving-cert\") pod \"controller-manager-5cf999dc48-9csf2\" (UID: \"18a6aea8-a3a2-4e08-848b-b33ead229eea\") " pod="openshift-controller-manager/controller-manager-5cf999dc48-9csf2" Oct 08 00:18:46 crc kubenswrapper[4967]: I1008 00:18:46.922258 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lw9xq\" (UniqueName: \"kubernetes.io/projected/18a6aea8-a3a2-4e08-848b-b33ead229eea-kube-api-access-lw9xq\") pod \"controller-manager-5cf999dc48-9csf2\" (UID: \"18a6aea8-a3a2-4e08-848b-b33ead229eea\") " pod="openshift-controller-manager/controller-manager-5cf999dc48-9csf2" Oct 08 00:18:47 crc kubenswrapper[4967]: I1008 00:18:47.109650 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5cf999dc48-9csf2" Oct 08 00:18:47 crc kubenswrapper[4967]: I1008 00:18:47.332549 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5cf999dc48-9csf2"] Oct 08 00:18:47 crc kubenswrapper[4967]: W1008 00:18:47.341467 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod18a6aea8_a3a2_4e08_848b_b33ead229eea.slice/crio-9d5f0afc6760c0f6cff7e25507baa20682a949bbd6903c58500abd79980a2406 WatchSource:0}: Error finding container 9d5f0afc6760c0f6cff7e25507baa20682a949bbd6903c58500abd79980a2406: Status 404 returned error can't find the container with id 9d5f0afc6760c0f6cff7e25507baa20682a949bbd6903c58500abd79980a2406 Oct 08 00:18:47 crc kubenswrapper[4967]: I1008 00:18:47.742385 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5cf999dc48-9csf2" event={"ID":"18a6aea8-a3a2-4e08-848b-b33ead229eea","Type":"ContainerStarted","Data":"9d5f0afc6760c0f6cff7e25507baa20682a949bbd6903c58500abd79980a2406"} Oct 08 00:18:48 crc kubenswrapper[4967]: I1008 00:18:48.126153 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8950261e-9b3f-443e-ab30-692f2604185a" path="/var/lib/kubelet/pods/8950261e-9b3f-443e-ab30-692f2604185a/volumes" Oct 08 00:18:48 crc kubenswrapper[4967]: I1008 00:18:48.127257 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d9e5f3be-2a1f-4890-acf7-6a9dae8f1f5a" path="/var/lib/kubelet/pods/d9e5f3be-2a1f-4890-acf7-6a9dae8f1f5a/volumes" Oct 08 00:18:48 crc kubenswrapper[4967]: I1008 00:18:48.749027 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5cf999dc48-9csf2" event={"ID":"18a6aea8-a3a2-4e08-848b-b33ead229eea","Type":"ContainerStarted","Data":"f68876e0646367744ce7168c7a97d44589791b086fdc5f3082a615ab5ac727d5"} Oct 08 00:18:48 crc kubenswrapper[4967]: I1008 00:18:48.749643 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-5cf999dc48-9csf2" Oct 08 00:18:48 crc kubenswrapper[4967]: I1008 00:18:48.755905 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-5cf999dc48-9csf2" Oct 08 00:18:48 crc kubenswrapper[4967]: I1008 00:18:48.766601 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-5cf999dc48-9csf2" podStartSLOduration=3.766584574 podStartE2EDuration="3.766584574s" podCreationTimestamp="2025-10-08 00:18:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:18:48.763437938 +0000 UTC m=+727.413972617" watchObservedRunningTime="2025-10-08 00:18:48.766584574 +0000 UTC m=+727.417119223" Oct 08 00:18:49 crc kubenswrapper[4967]: I1008 00:18:49.435499 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-58c65d6798-7drt6"] Oct 08 00:18:49 crc kubenswrapper[4967]: I1008 00:18:49.440038 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-58c65d6798-7drt6" Oct 08 00:18:49 crc kubenswrapper[4967]: I1008 00:18:49.443293 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 08 00:18:49 crc kubenswrapper[4967]: I1008 00:18:49.447503 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-58c65d6798-7drt6"] Oct 08 00:18:49 crc kubenswrapper[4967]: I1008 00:18:49.448430 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 08 00:18:49 crc kubenswrapper[4967]: I1008 00:18:49.451971 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 08 00:18:49 crc kubenswrapper[4967]: I1008 00:18:49.452216 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 08 00:18:49 crc kubenswrapper[4967]: I1008 00:18:49.452366 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 08 00:18:49 crc kubenswrapper[4967]: I1008 00:18:49.452217 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 08 00:18:49 crc kubenswrapper[4967]: I1008 00:18:49.523119 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b1bdb6cd-5707-4ef4-8588-f6a8e76c4f1f-config\") pod \"route-controller-manager-58c65d6798-7drt6\" (UID: \"b1bdb6cd-5707-4ef4-8588-f6a8e76c4f1f\") " pod="openshift-route-controller-manager/route-controller-manager-58c65d6798-7drt6" Oct 08 00:18:49 crc kubenswrapper[4967]: I1008 00:18:49.523380 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b1bdb6cd-5707-4ef4-8588-f6a8e76c4f1f-serving-cert\") pod \"route-controller-manager-58c65d6798-7drt6\" (UID: \"b1bdb6cd-5707-4ef4-8588-f6a8e76c4f1f\") " pod="openshift-route-controller-manager/route-controller-manager-58c65d6798-7drt6" Oct 08 00:18:49 crc kubenswrapper[4967]: I1008 00:18:49.523422 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b1bdb6cd-5707-4ef4-8588-f6a8e76c4f1f-client-ca\") pod \"route-controller-manager-58c65d6798-7drt6\" (UID: \"b1bdb6cd-5707-4ef4-8588-f6a8e76c4f1f\") " pod="openshift-route-controller-manager/route-controller-manager-58c65d6798-7drt6" Oct 08 00:18:49 crc kubenswrapper[4967]: I1008 00:18:49.523443 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cgnv6\" (UniqueName: \"kubernetes.io/projected/b1bdb6cd-5707-4ef4-8588-f6a8e76c4f1f-kube-api-access-cgnv6\") pod \"route-controller-manager-58c65d6798-7drt6\" (UID: \"b1bdb6cd-5707-4ef4-8588-f6a8e76c4f1f\") " pod="openshift-route-controller-manager/route-controller-manager-58c65d6798-7drt6" Oct 08 00:18:49 crc kubenswrapper[4967]: I1008 00:18:49.624527 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b1bdb6cd-5707-4ef4-8588-f6a8e76c4f1f-client-ca\") pod \"route-controller-manager-58c65d6798-7drt6\" (UID: \"b1bdb6cd-5707-4ef4-8588-f6a8e76c4f1f\") " pod="openshift-route-controller-manager/route-controller-manager-58c65d6798-7drt6" Oct 08 00:18:49 crc kubenswrapper[4967]: I1008 00:18:49.624572 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cgnv6\" (UniqueName: \"kubernetes.io/projected/b1bdb6cd-5707-4ef4-8588-f6a8e76c4f1f-kube-api-access-cgnv6\") pod \"route-controller-manager-58c65d6798-7drt6\" (UID: \"b1bdb6cd-5707-4ef4-8588-f6a8e76c4f1f\") " pod="openshift-route-controller-manager/route-controller-manager-58c65d6798-7drt6" Oct 08 00:18:49 crc kubenswrapper[4967]: I1008 00:18:49.624624 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b1bdb6cd-5707-4ef4-8588-f6a8e76c4f1f-config\") pod \"route-controller-manager-58c65d6798-7drt6\" (UID: \"b1bdb6cd-5707-4ef4-8588-f6a8e76c4f1f\") " pod="openshift-route-controller-manager/route-controller-manager-58c65d6798-7drt6" Oct 08 00:18:49 crc kubenswrapper[4967]: I1008 00:18:49.624640 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b1bdb6cd-5707-4ef4-8588-f6a8e76c4f1f-serving-cert\") pod \"route-controller-manager-58c65d6798-7drt6\" (UID: \"b1bdb6cd-5707-4ef4-8588-f6a8e76c4f1f\") " pod="openshift-route-controller-manager/route-controller-manager-58c65d6798-7drt6" Oct 08 00:18:49 crc kubenswrapper[4967]: I1008 00:18:49.626268 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b1bdb6cd-5707-4ef4-8588-f6a8e76c4f1f-client-ca\") pod \"route-controller-manager-58c65d6798-7drt6\" (UID: \"b1bdb6cd-5707-4ef4-8588-f6a8e76c4f1f\") " pod="openshift-route-controller-manager/route-controller-manager-58c65d6798-7drt6" Oct 08 00:18:49 crc kubenswrapper[4967]: I1008 00:18:49.626533 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b1bdb6cd-5707-4ef4-8588-f6a8e76c4f1f-config\") pod \"route-controller-manager-58c65d6798-7drt6\" (UID: \"b1bdb6cd-5707-4ef4-8588-f6a8e76c4f1f\") " pod="openshift-route-controller-manager/route-controller-manager-58c65d6798-7drt6" Oct 08 00:18:49 crc kubenswrapper[4967]: I1008 00:18:49.633371 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b1bdb6cd-5707-4ef4-8588-f6a8e76c4f1f-serving-cert\") pod \"route-controller-manager-58c65d6798-7drt6\" (UID: \"b1bdb6cd-5707-4ef4-8588-f6a8e76c4f1f\") " pod="openshift-route-controller-manager/route-controller-manager-58c65d6798-7drt6" Oct 08 00:18:49 crc kubenswrapper[4967]: I1008 00:18:49.642324 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cgnv6\" (UniqueName: \"kubernetes.io/projected/b1bdb6cd-5707-4ef4-8588-f6a8e76c4f1f-kube-api-access-cgnv6\") pod \"route-controller-manager-58c65d6798-7drt6\" (UID: \"b1bdb6cd-5707-4ef4-8588-f6a8e76c4f1f\") " pod="openshift-route-controller-manager/route-controller-manager-58c65d6798-7drt6" Oct 08 00:18:49 crc kubenswrapper[4967]: I1008 00:18:49.755005 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-58c65d6798-7drt6" Oct 08 00:18:50 crc kubenswrapper[4967]: I1008 00:18:50.186157 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-58c65d6798-7drt6"] Oct 08 00:18:50 crc kubenswrapper[4967]: W1008 00:18:50.192123 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb1bdb6cd_5707_4ef4_8588_f6a8e76c4f1f.slice/crio-a58c85be6321fb49895f1234b4c300660183937b78fcd3ea7e06e25b0aae0257 WatchSource:0}: Error finding container a58c85be6321fb49895f1234b4c300660183937b78fcd3ea7e06e25b0aae0257: Status 404 returned error can't find the container with id a58c85be6321fb49895f1234b4c300660183937b78fcd3ea7e06e25b0aae0257 Oct 08 00:18:50 crc kubenswrapper[4967]: I1008 00:18:50.760851 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-58c65d6798-7drt6" event={"ID":"b1bdb6cd-5707-4ef4-8588-f6a8e76c4f1f","Type":"ContainerStarted","Data":"ea207e6660a47f0cee26fb7ed77df9eb7c589c6a500088be0fa5f8e65dc938ea"} Oct 08 00:18:50 crc kubenswrapper[4967]: I1008 00:18:50.761323 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-58c65d6798-7drt6" event={"ID":"b1bdb6cd-5707-4ef4-8588-f6a8e76c4f1f","Type":"ContainerStarted","Data":"a58c85be6321fb49895f1234b4c300660183937b78fcd3ea7e06e25b0aae0257"} Oct 08 00:18:50 crc kubenswrapper[4967]: I1008 00:18:50.777044 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-58c65d6798-7drt6" podStartSLOduration=5.777023596 podStartE2EDuration="5.777023596s" podCreationTimestamp="2025-10-08 00:18:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:18:50.774610439 +0000 UTC m=+729.425145088" watchObservedRunningTime="2025-10-08 00:18:50.777023596 +0000 UTC m=+729.427558245" Oct 08 00:18:51 crc kubenswrapper[4967]: I1008 00:18:51.766217 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-58c65d6798-7drt6" Oct 08 00:18:51 crc kubenswrapper[4967]: I1008 00:18:51.771190 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-58c65d6798-7drt6" Oct 08 00:18:55 crc kubenswrapper[4967]: I1008 00:18:55.236161 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-hrfsc"] Oct 08 00:18:55 crc kubenswrapper[4967]: I1008 00:18:55.237904 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hrfsc" Oct 08 00:18:55 crc kubenswrapper[4967]: I1008 00:18:55.248992 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hrfsc"] Oct 08 00:18:55 crc kubenswrapper[4967]: I1008 00:18:55.319526 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b8900b3-7e99-48c8-a5a9-52a6edd4f7e1-utilities\") pod \"community-operators-hrfsc\" (UID: \"0b8900b3-7e99-48c8-a5a9-52a6edd4f7e1\") " pod="openshift-marketplace/community-operators-hrfsc" Oct 08 00:18:55 crc kubenswrapper[4967]: I1008 00:18:55.319612 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pbd6f\" (UniqueName: \"kubernetes.io/projected/0b8900b3-7e99-48c8-a5a9-52a6edd4f7e1-kube-api-access-pbd6f\") pod \"community-operators-hrfsc\" (UID: \"0b8900b3-7e99-48c8-a5a9-52a6edd4f7e1\") " pod="openshift-marketplace/community-operators-hrfsc" Oct 08 00:18:55 crc kubenswrapper[4967]: I1008 00:18:55.319646 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b8900b3-7e99-48c8-a5a9-52a6edd4f7e1-catalog-content\") pod \"community-operators-hrfsc\" (UID: \"0b8900b3-7e99-48c8-a5a9-52a6edd4f7e1\") " pod="openshift-marketplace/community-operators-hrfsc" Oct 08 00:18:55 crc kubenswrapper[4967]: I1008 00:18:55.421273 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b8900b3-7e99-48c8-a5a9-52a6edd4f7e1-utilities\") pod \"community-operators-hrfsc\" (UID: \"0b8900b3-7e99-48c8-a5a9-52a6edd4f7e1\") " pod="openshift-marketplace/community-operators-hrfsc" Oct 08 00:18:55 crc kubenswrapper[4967]: I1008 00:18:55.421358 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pbd6f\" (UniqueName: \"kubernetes.io/projected/0b8900b3-7e99-48c8-a5a9-52a6edd4f7e1-kube-api-access-pbd6f\") pod \"community-operators-hrfsc\" (UID: \"0b8900b3-7e99-48c8-a5a9-52a6edd4f7e1\") " pod="openshift-marketplace/community-operators-hrfsc" Oct 08 00:18:55 crc kubenswrapper[4967]: I1008 00:18:55.421388 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b8900b3-7e99-48c8-a5a9-52a6edd4f7e1-catalog-content\") pod \"community-operators-hrfsc\" (UID: \"0b8900b3-7e99-48c8-a5a9-52a6edd4f7e1\") " pod="openshift-marketplace/community-operators-hrfsc" Oct 08 00:18:55 crc kubenswrapper[4967]: I1008 00:18:55.421976 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b8900b3-7e99-48c8-a5a9-52a6edd4f7e1-catalog-content\") pod \"community-operators-hrfsc\" (UID: \"0b8900b3-7e99-48c8-a5a9-52a6edd4f7e1\") " pod="openshift-marketplace/community-operators-hrfsc" Oct 08 00:18:55 crc kubenswrapper[4967]: I1008 00:18:55.422015 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b8900b3-7e99-48c8-a5a9-52a6edd4f7e1-utilities\") pod \"community-operators-hrfsc\" (UID: \"0b8900b3-7e99-48c8-a5a9-52a6edd4f7e1\") " pod="openshift-marketplace/community-operators-hrfsc" Oct 08 00:18:55 crc kubenswrapper[4967]: I1008 00:18:55.441804 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pbd6f\" (UniqueName: \"kubernetes.io/projected/0b8900b3-7e99-48c8-a5a9-52a6edd4f7e1-kube-api-access-pbd6f\") pod \"community-operators-hrfsc\" (UID: \"0b8900b3-7e99-48c8-a5a9-52a6edd4f7e1\") " pod="openshift-marketplace/community-operators-hrfsc" Oct 08 00:18:55 crc kubenswrapper[4967]: I1008 00:18:55.573780 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hrfsc" Oct 08 00:18:56 crc kubenswrapper[4967]: I1008 00:18:56.045892 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hrfsc"] Oct 08 00:18:56 crc kubenswrapper[4967]: W1008 00:18:56.049984 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0b8900b3_7e99_48c8_a5a9_52a6edd4f7e1.slice/crio-096121531ddfdd01595f02697c0732054f478d352cfa7eef1e24a6a78d075f70 WatchSource:0}: Error finding container 096121531ddfdd01595f02697c0732054f478d352cfa7eef1e24a6a78d075f70: Status 404 returned error can't find the container with id 096121531ddfdd01595f02697c0732054f478d352cfa7eef1e24a6a78d075f70 Oct 08 00:18:56 crc kubenswrapper[4967]: I1008 00:18:56.823512 4967 generic.go:334] "Generic (PLEG): container finished" podID="0b8900b3-7e99-48c8-a5a9-52a6edd4f7e1" containerID="210414d88dd93e31f05a5c086c7b476c52c47263c4d00a05ed0605da16b277e8" exitCode=0 Oct 08 00:18:56 crc kubenswrapper[4967]: I1008 00:18:56.823586 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hrfsc" event={"ID":"0b8900b3-7e99-48c8-a5a9-52a6edd4f7e1","Type":"ContainerDied","Data":"210414d88dd93e31f05a5c086c7b476c52c47263c4d00a05ed0605da16b277e8"} Oct 08 00:18:56 crc kubenswrapper[4967]: I1008 00:18:56.824650 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hrfsc" event={"ID":"0b8900b3-7e99-48c8-a5a9-52a6edd4f7e1","Type":"ContainerStarted","Data":"096121531ddfdd01595f02697c0732054f478d352cfa7eef1e24a6a78d075f70"} Oct 08 00:18:57 crc kubenswrapper[4967]: I1008 00:18:57.011860 4967 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 08 00:18:58 crc kubenswrapper[4967]: I1008 00:18:58.849072 4967 generic.go:334] "Generic (PLEG): container finished" podID="0b8900b3-7e99-48c8-a5a9-52a6edd4f7e1" containerID="38e6993736fec464ea8e9209b3e8cd4da3eb6db74a4b831be9b64393af385f22" exitCode=0 Oct 08 00:18:58 crc kubenswrapper[4967]: I1008 00:18:58.849121 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hrfsc" event={"ID":"0b8900b3-7e99-48c8-a5a9-52a6edd4f7e1","Type":"ContainerDied","Data":"38e6993736fec464ea8e9209b3e8cd4da3eb6db74a4b831be9b64393af385f22"} Oct 08 00:18:59 crc kubenswrapper[4967]: I1008 00:18:59.859242 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hrfsc" event={"ID":"0b8900b3-7e99-48c8-a5a9-52a6edd4f7e1","Type":"ContainerStarted","Data":"45674a976f6cb481ed3eab9e7f64f0e2a2b2c19cbea91862e4b4b804d2a12cc9"} Oct 08 00:18:59 crc kubenswrapper[4967]: I1008 00:18:59.880919 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-hrfsc" podStartSLOduration=2.397796029 podStartE2EDuration="4.880893933s" podCreationTimestamp="2025-10-08 00:18:55 +0000 UTC" firstStartedPulling="2025-10-08 00:18:56.826143008 +0000 UTC m=+735.476677667" lastFinishedPulling="2025-10-08 00:18:59.309240912 +0000 UTC m=+737.959775571" observedRunningTime="2025-10-08 00:18:59.878997841 +0000 UTC m=+738.529532490" watchObservedRunningTime="2025-10-08 00:18:59.880893933 +0000 UTC m=+738.531428602" Oct 08 00:19:04 crc kubenswrapper[4967]: I1008 00:19:04.660430 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-xmr9q"] Oct 08 00:19:04 crc kubenswrapper[4967]: I1008 00:19:04.662047 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xmr9q" Oct 08 00:19:04 crc kubenswrapper[4967]: I1008 00:19:04.680765 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xmr9q"] Oct 08 00:19:04 crc kubenswrapper[4967]: I1008 00:19:04.781523 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3ad3b862-dba5-46ed-b610-2c95c42183ce-utilities\") pod \"certified-operators-xmr9q\" (UID: \"3ad3b862-dba5-46ed-b610-2c95c42183ce\") " pod="openshift-marketplace/certified-operators-xmr9q" Oct 08 00:19:04 crc kubenswrapper[4967]: I1008 00:19:04.781607 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r2b7m\" (UniqueName: \"kubernetes.io/projected/3ad3b862-dba5-46ed-b610-2c95c42183ce-kube-api-access-r2b7m\") pod \"certified-operators-xmr9q\" (UID: \"3ad3b862-dba5-46ed-b610-2c95c42183ce\") " pod="openshift-marketplace/certified-operators-xmr9q" Oct 08 00:19:04 crc kubenswrapper[4967]: I1008 00:19:04.781639 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3ad3b862-dba5-46ed-b610-2c95c42183ce-catalog-content\") pod \"certified-operators-xmr9q\" (UID: \"3ad3b862-dba5-46ed-b610-2c95c42183ce\") " pod="openshift-marketplace/certified-operators-xmr9q" Oct 08 00:19:04 crc kubenswrapper[4967]: I1008 00:19:04.882925 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3ad3b862-dba5-46ed-b610-2c95c42183ce-catalog-content\") pod \"certified-operators-xmr9q\" (UID: \"3ad3b862-dba5-46ed-b610-2c95c42183ce\") " pod="openshift-marketplace/certified-operators-xmr9q" Oct 08 00:19:04 crc kubenswrapper[4967]: I1008 00:19:04.883544 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3ad3b862-dba5-46ed-b610-2c95c42183ce-utilities\") pod \"certified-operators-xmr9q\" (UID: \"3ad3b862-dba5-46ed-b610-2c95c42183ce\") " pod="openshift-marketplace/certified-operators-xmr9q" Oct 08 00:19:04 crc kubenswrapper[4967]: I1008 00:19:04.883923 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r2b7m\" (UniqueName: \"kubernetes.io/projected/3ad3b862-dba5-46ed-b610-2c95c42183ce-kube-api-access-r2b7m\") pod \"certified-operators-xmr9q\" (UID: \"3ad3b862-dba5-46ed-b610-2c95c42183ce\") " pod="openshift-marketplace/certified-operators-xmr9q" Oct 08 00:19:04 crc kubenswrapper[4967]: I1008 00:19:04.883848 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3ad3b862-dba5-46ed-b610-2c95c42183ce-utilities\") pod \"certified-operators-xmr9q\" (UID: \"3ad3b862-dba5-46ed-b610-2c95c42183ce\") " pod="openshift-marketplace/certified-operators-xmr9q" Oct 08 00:19:04 crc kubenswrapper[4967]: I1008 00:19:04.883491 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3ad3b862-dba5-46ed-b610-2c95c42183ce-catalog-content\") pod \"certified-operators-xmr9q\" (UID: \"3ad3b862-dba5-46ed-b610-2c95c42183ce\") " pod="openshift-marketplace/certified-operators-xmr9q" Oct 08 00:19:04 crc kubenswrapper[4967]: I1008 00:19:04.905502 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r2b7m\" (UniqueName: \"kubernetes.io/projected/3ad3b862-dba5-46ed-b610-2c95c42183ce-kube-api-access-r2b7m\") pod \"certified-operators-xmr9q\" (UID: \"3ad3b862-dba5-46ed-b610-2c95c42183ce\") " pod="openshift-marketplace/certified-operators-xmr9q" Oct 08 00:19:05 crc kubenswrapper[4967]: I1008 00:19:05.007212 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xmr9q" Oct 08 00:19:05 crc kubenswrapper[4967]: I1008 00:19:05.504385 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xmr9q"] Oct 08 00:19:05 crc kubenswrapper[4967]: I1008 00:19:05.574562 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-hrfsc" Oct 08 00:19:05 crc kubenswrapper[4967]: I1008 00:19:05.574604 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-hrfsc" Oct 08 00:19:05 crc kubenswrapper[4967]: I1008 00:19:05.618353 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-hrfsc" Oct 08 00:19:05 crc kubenswrapper[4967]: I1008 00:19:05.903891 4967 generic.go:334] "Generic (PLEG): container finished" podID="3ad3b862-dba5-46ed-b610-2c95c42183ce" containerID="fbf2abb1ff1e3da0257fca13f622aadfb81e62a680d30b2f309fd7c5c118d8df" exitCode=0 Oct 08 00:19:05 crc kubenswrapper[4967]: I1008 00:19:05.904040 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xmr9q" event={"ID":"3ad3b862-dba5-46ed-b610-2c95c42183ce","Type":"ContainerDied","Data":"fbf2abb1ff1e3da0257fca13f622aadfb81e62a680d30b2f309fd7c5c118d8df"} Oct 08 00:19:05 crc kubenswrapper[4967]: I1008 00:19:05.904086 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xmr9q" event={"ID":"3ad3b862-dba5-46ed-b610-2c95c42183ce","Type":"ContainerStarted","Data":"39c99e185f9f446e95660a4f125cdf3c70f74077d2965658711a37262f11a1aa"} Oct 08 00:19:05 crc kubenswrapper[4967]: I1008 00:19:05.963206 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-hrfsc" Oct 08 00:19:06 crc kubenswrapper[4967]: I1008 00:19:06.917011 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xmr9q" event={"ID":"3ad3b862-dba5-46ed-b610-2c95c42183ce","Type":"ContainerStarted","Data":"5ccb1d8d4b9aefd38e47e8bf8b33cca404fb67b7b75b6002c8adfa356c52fb5d"} Oct 08 00:19:07 crc kubenswrapper[4967]: I1008 00:19:07.924295 4967 generic.go:334] "Generic (PLEG): container finished" podID="3ad3b862-dba5-46ed-b610-2c95c42183ce" containerID="5ccb1d8d4b9aefd38e47e8bf8b33cca404fb67b7b75b6002c8adfa356c52fb5d" exitCode=0 Oct 08 00:19:07 crc kubenswrapper[4967]: I1008 00:19:07.924539 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xmr9q" event={"ID":"3ad3b862-dba5-46ed-b610-2c95c42183ce","Type":"ContainerDied","Data":"5ccb1d8d4b9aefd38e47e8bf8b33cca404fb67b7b75b6002c8adfa356c52fb5d"} Oct 08 00:19:08 crc kubenswrapper[4967]: I1008 00:19:08.036754 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hrfsc"] Oct 08 00:19:08 crc kubenswrapper[4967]: I1008 00:19:08.037085 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-hrfsc" podUID="0b8900b3-7e99-48c8-a5a9-52a6edd4f7e1" containerName="registry-server" containerID="cri-o://45674a976f6cb481ed3eab9e7f64f0e2a2b2c19cbea91862e4b4b804d2a12cc9" gracePeriod=2 Oct 08 00:19:08 crc kubenswrapper[4967]: I1008 00:19:08.933559 4967 generic.go:334] "Generic (PLEG): container finished" podID="0b8900b3-7e99-48c8-a5a9-52a6edd4f7e1" containerID="45674a976f6cb481ed3eab9e7f64f0e2a2b2c19cbea91862e4b4b804d2a12cc9" exitCode=0 Oct 08 00:19:08 crc kubenswrapper[4967]: I1008 00:19:08.933685 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hrfsc" event={"ID":"0b8900b3-7e99-48c8-a5a9-52a6edd4f7e1","Type":"ContainerDied","Data":"45674a976f6cb481ed3eab9e7f64f0e2a2b2c19cbea91862e4b4b804d2a12cc9"} Oct 08 00:19:09 crc kubenswrapper[4967]: I1008 00:19:09.119267 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hrfsc" Oct 08 00:19:09 crc kubenswrapper[4967]: I1008 00:19:09.241701 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pbd6f\" (UniqueName: \"kubernetes.io/projected/0b8900b3-7e99-48c8-a5a9-52a6edd4f7e1-kube-api-access-pbd6f\") pod \"0b8900b3-7e99-48c8-a5a9-52a6edd4f7e1\" (UID: \"0b8900b3-7e99-48c8-a5a9-52a6edd4f7e1\") " Oct 08 00:19:09 crc kubenswrapper[4967]: I1008 00:19:09.241866 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b8900b3-7e99-48c8-a5a9-52a6edd4f7e1-utilities\") pod \"0b8900b3-7e99-48c8-a5a9-52a6edd4f7e1\" (UID: \"0b8900b3-7e99-48c8-a5a9-52a6edd4f7e1\") " Oct 08 00:19:09 crc kubenswrapper[4967]: I1008 00:19:09.241915 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b8900b3-7e99-48c8-a5a9-52a6edd4f7e1-catalog-content\") pod \"0b8900b3-7e99-48c8-a5a9-52a6edd4f7e1\" (UID: \"0b8900b3-7e99-48c8-a5a9-52a6edd4f7e1\") " Oct 08 00:19:09 crc kubenswrapper[4967]: I1008 00:19:09.243092 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0b8900b3-7e99-48c8-a5a9-52a6edd4f7e1-utilities" (OuterVolumeSpecName: "utilities") pod "0b8900b3-7e99-48c8-a5a9-52a6edd4f7e1" (UID: "0b8900b3-7e99-48c8-a5a9-52a6edd4f7e1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:19:09 crc kubenswrapper[4967]: I1008 00:19:09.250904 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b8900b3-7e99-48c8-a5a9-52a6edd4f7e1-kube-api-access-pbd6f" (OuterVolumeSpecName: "kube-api-access-pbd6f") pod "0b8900b3-7e99-48c8-a5a9-52a6edd4f7e1" (UID: "0b8900b3-7e99-48c8-a5a9-52a6edd4f7e1"). InnerVolumeSpecName "kube-api-access-pbd6f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:19:09 crc kubenswrapper[4967]: I1008 00:19:09.287762 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0b8900b3-7e99-48c8-a5a9-52a6edd4f7e1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0b8900b3-7e99-48c8-a5a9-52a6edd4f7e1" (UID: "0b8900b3-7e99-48c8-a5a9-52a6edd4f7e1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:19:09 crc kubenswrapper[4967]: I1008 00:19:09.343118 4967 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b8900b3-7e99-48c8-a5a9-52a6edd4f7e1-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 00:19:09 crc kubenswrapper[4967]: I1008 00:19:09.343337 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pbd6f\" (UniqueName: \"kubernetes.io/projected/0b8900b3-7e99-48c8-a5a9-52a6edd4f7e1-kube-api-access-pbd6f\") on node \"crc\" DevicePath \"\"" Oct 08 00:19:09 crc kubenswrapper[4967]: I1008 00:19:09.343406 4967 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b8900b3-7e99-48c8-a5a9-52a6edd4f7e1-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 00:19:09 crc kubenswrapper[4967]: I1008 00:19:09.943151 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hrfsc" event={"ID":"0b8900b3-7e99-48c8-a5a9-52a6edd4f7e1","Type":"ContainerDied","Data":"096121531ddfdd01595f02697c0732054f478d352cfa7eef1e24a6a78d075f70"} Oct 08 00:19:09 crc kubenswrapper[4967]: I1008 00:19:09.943180 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hrfsc" Oct 08 00:19:09 crc kubenswrapper[4967]: I1008 00:19:09.944354 4967 scope.go:117] "RemoveContainer" containerID="45674a976f6cb481ed3eab9e7f64f0e2a2b2c19cbea91862e4b4b804d2a12cc9" Oct 08 00:19:09 crc kubenswrapper[4967]: I1008 00:19:09.945299 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xmr9q" event={"ID":"3ad3b862-dba5-46ed-b610-2c95c42183ce","Type":"ContainerStarted","Data":"7939fb05cc0c8802da6c333fdd5eae5113cf4d3f69cbe4356bda8f0755e260e4"} Oct 08 00:19:09 crc kubenswrapper[4967]: I1008 00:19:09.963874 4967 scope.go:117] "RemoveContainer" containerID="38e6993736fec464ea8e9209b3e8cd4da3eb6db74a4b831be9b64393af385f22" Oct 08 00:19:09 crc kubenswrapper[4967]: I1008 00:19:09.969267 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-xmr9q" podStartSLOduration=2.861900624 podStartE2EDuration="5.969249363s" podCreationTimestamp="2025-10-08 00:19:04 +0000 UTC" firstStartedPulling="2025-10-08 00:19:05.905478592 +0000 UTC m=+744.556013241" lastFinishedPulling="2025-10-08 00:19:09.012827331 +0000 UTC m=+747.663361980" observedRunningTime="2025-10-08 00:19:09.965477948 +0000 UTC m=+748.616012617" watchObservedRunningTime="2025-10-08 00:19:09.969249363 +0000 UTC m=+748.619784002" Oct 08 00:19:09 crc kubenswrapper[4967]: I1008 00:19:09.982718 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hrfsc"] Oct 08 00:19:09 crc kubenswrapper[4967]: I1008 00:19:09.990520 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-hrfsc"] Oct 08 00:19:09 crc kubenswrapper[4967]: I1008 00:19:09.992659 4967 scope.go:117] "RemoveContainer" containerID="210414d88dd93e31f05a5c086c7b476c52c47263c4d00a05ed0605da16b277e8" Oct 08 00:19:10 crc kubenswrapper[4967]: I1008 00:19:10.125578 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b8900b3-7e99-48c8-a5a9-52a6edd4f7e1" path="/var/lib/kubelet/pods/0b8900b3-7e99-48c8-a5a9-52a6edd4f7e1/volumes" Oct 08 00:19:15 crc kubenswrapper[4967]: I1008 00:19:15.007515 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-xmr9q" Oct 08 00:19:15 crc kubenswrapper[4967]: I1008 00:19:15.008223 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-xmr9q" Oct 08 00:19:15 crc kubenswrapper[4967]: I1008 00:19:15.060089 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-xmr9q" Oct 08 00:19:16 crc kubenswrapper[4967]: I1008 00:19:16.066015 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-xmr9q" Oct 08 00:19:16 crc kubenswrapper[4967]: I1008 00:19:16.140095 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xmr9q"] Oct 08 00:19:18 crc kubenswrapper[4967]: I1008 00:19:18.002076 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-xmr9q" podUID="3ad3b862-dba5-46ed-b610-2c95c42183ce" containerName="registry-server" containerID="cri-o://7939fb05cc0c8802da6c333fdd5eae5113cf4d3f69cbe4356bda8f0755e260e4" gracePeriod=2 Oct 08 00:19:19 crc kubenswrapper[4967]: I1008 00:19:19.017053 4967 generic.go:334] "Generic (PLEG): container finished" podID="3ad3b862-dba5-46ed-b610-2c95c42183ce" containerID="7939fb05cc0c8802da6c333fdd5eae5113cf4d3f69cbe4356bda8f0755e260e4" exitCode=0 Oct 08 00:19:19 crc kubenswrapper[4967]: I1008 00:19:19.018429 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xmr9q" event={"ID":"3ad3b862-dba5-46ed-b610-2c95c42183ce","Type":"ContainerDied","Data":"7939fb05cc0c8802da6c333fdd5eae5113cf4d3f69cbe4356bda8f0755e260e4"} Oct 08 00:19:19 crc kubenswrapper[4967]: I1008 00:19:19.075331 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xmr9q" Oct 08 00:19:19 crc kubenswrapper[4967]: I1008 00:19:19.186094 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3ad3b862-dba5-46ed-b610-2c95c42183ce-catalog-content\") pod \"3ad3b862-dba5-46ed-b610-2c95c42183ce\" (UID: \"3ad3b862-dba5-46ed-b610-2c95c42183ce\") " Oct 08 00:19:19 crc kubenswrapper[4967]: I1008 00:19:19.186181 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r2b7m\" (UniqueName: \"kubernetes.io/projected/3ad3b862-dba5-46ed-b610-2c95c42183ce-kube-api-access-r2b7m\") pod \"3ad3b862-dba5-46ed-b610-2c95c42183ce\" (UID: \"3ad3b862-dba5-46ed-b610-2c95c42183ce\") " Oct 08 00:19:19 crc kubenswrapper[4967]: I1008 00:19:19.186213 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3ad3b862-dba5-46ed-b610-2c95c42183ce-utilities\") pod \"3ad3b862-dba5-46ed-b610-2c95c42183ce\" (UID: \"3ad3b862-dba5-46ed-b610-2c95c42183ce\") " Oct 08 00:19:19 crc kubenswrapper[4967]: I1008 00:19:19.188204 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3ad3b862-dba5-46ed-b610-2c95c42183ce-utilities" (OuterVolumeSpecName: "utilities") pod "3ad3b862-dba5-46ed-b610-2c95c42183ce" (UID: "3ad3b862-dba5-46ed-b610-2c95c42183ce"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:19:19 crc kubenswrapper[4967]: I1008 00:19:19.194498 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ad3b862-dba5-46ed-b610-2c95c42183ce-kube-api-access-r2b7m" (OuterVolumeSpecName: "kube-api-access-r2b7m") pod "3ad3b862-dba5-46ed-b610-2c95c42183ce" (UID: "3ad3b862-dba5-46ed-b610-2c95c42183ce"). InnerVolumeSpecName "kube-api-access-r2b7m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:19:19 crc kubenswrapper[4967]: I1008 00:19:19.260434 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3ad3b862-dba5-46ed-b610-2c95c42183ce-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3ad3b862-dba5-46ed-b610-2c95c42183ce" (UID: "3ad3b862-dba5-46ed-b610-2c95c42183ce"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:19:19 crc kubenswrapper[4967]: I1008 00:19:19.299776 4967 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3ad3b862-dba5-46ed-b610-2c95c42183ce-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 00:19:19 crc kubenswrapper[4967]: I1008 00:19:19.299835 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r2b7m\" (UniqueName: \"kubernetes.io/projected/3ad3b862-dba5-46ed-b610-2c95c42183ce-kube-api-access-r2b7m\") on node \"crc\" DevicePath \"\"" Oct 08 00:19:19 crc kubenswrapper[4967]: I1008 00:19:19.299861 4967 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3ad3b862-dba5-46ed-b610-2c95c42183ce-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 00:19:20 crc kubenswrapper[4967]: I1008 00:19:20.031077 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xmr9q" event={"ID":"3ad3b862-dba5-46ed-b610-2c95c42183ce","Type":"ContainerDied","Data":"39c99e185f9f446e95660a4f125cdf3c70f74077d2965658711a37262f11a1aa"} Oct 08 00:19:20 crc kubenswrapper[4967]: I1008 00:19:20.031176 4967 scope.go:117] "RemoveContainer" containerID="7939fb05cc0c8802da6c333fdd5eae5113cf4d3f69cbe4356bda8f0755e260e4" Oct 08 00:19:20 crc kubenswrapper[4967]: I1008 00:19:20.031182 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xmr9q" Oct 08 00:19:20 crc kubenswrapper[4967]: I1008 00:19:20.070778 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xmr9q"] Oct 08 00:19:20 crc kubenswrapper[4967]: I1008 00:19:20.075575 4967 scope.go:117] "RemoveContainer" containerID="5ccb1d8d4b9aefd38e47e8bf8b33cca404fb67b7b75b6002c8adfa356c52fb5d" Oct 08 00:19:20 crc kubenswrapper[4967]: I1008 00:19:20.076304 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-xmr9q"] Oct 08 00:19:20 crc kubenswrapper[4967]: I1008 00:19:20.100507 4967 scope.go:117] "RemoveContainer" containerID="fbf2abb1ff1e3da0257fca13f622aadfb81e62a680d30b2f309fd7c5c118d8df" Oct 08 00:19:20 crc kubenswrapper[4967]: I1008 00:19:20.126603 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ad3b862-dba5-46ed-b610-2c95c42183ce" path="/var/lib/kubelet/pods/3ad3b862-dba5-46ed-b610-2c95c42183ce/volumes" Oct 08 00:19:25 crc kubenswrapper[4967]: I1008 00:19:25.148588 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-nns67"] Oct 08 00:19:25 crc kubenswrapper[4967]: E1008 00:19:25.151465 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b8900b3-7e99-48c8-a5a9-52a6edd4f7e1" containerName="registry-server" Oct 08 00:19:25 crc kubenswrapper[4967]: I1008 00:19:25.151636 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b8900b3-7e99-48c8-a5a9-52a6edd4f7e1" containerName="registry-server" Oct 08 00:19:25 crc kubenswrapper[4967]: E1008 00:19:25.151804 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ad3b862-dba5-46ed-b610-2c95c42183ce" containerName="extract-content" Oct 08 00:19:25 crc kubenswrapper[4967]: I1008 00:19:25.151877 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ad3b862-dba5-46ed-b610-2c95c42183ce" containerName="extract-content" Oct 08 00:19:25 crc kubenswrapper[4967]: E1008 00:19:25.151951 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b8900b3-7e99-48c8-a5a9-52a6edd4f7e1" containerName="extract-content" Oct 08 00:19:25 crc kubenswrapper[4967]: I1008 00:19:25.152024 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b8900b3-7e99-48c8-a5a9-52a6edd4f7e1" containerName="extract-content" Oct 08 00:19:25 crc kubenswrapper[4967]: E1008 00:19:25.152093 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b8900b3-7e99-48c8-a5a9-52a6edd4f7e1" containerName="extract-utilities" Oct 08 00:19:25 crc kubenswrapper[4967]: I1008 00:19:25.152158 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b8900b3-7e99-48c8-a5a9-52a6edd4f7e1" containerName="extract-utilities" Oct 08 00:19:25 crc kubenswrapper[4967]: E1008 00:19:25.152248 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ad3b862-dba5-46ed-b610-2c95c42183ce" containerName="extract-utilities" Oct 08 00:19:25 crc kubenswrapper[4967]: I1008 00:19:25.152328 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ad3b862-dba5-46ed-b610-2c95c42183ce" containerName="extract-utilities" Oct 08 00:19:25 crc kubenswrapper[4967]: E1008 00:19:25.152404 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ad3b862-dba5-46ed-b610-2c95c42183ce" containerName="registry-server" Oct 08 00:19:25 crc kubenswrapper[4967]: I1008 00:19:25.152470 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ad3b862-dba5-46ed-b610-2c95c42183ce" containerName="registry-server" Oct 08 00:19:25 crc kubenswrapper[4967]: I1008 00:19:25.152696 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ad3b862-dba5-46ed-b610-2c95c42183ce" containerName="registry-server" Oct 08 00:19:25 crc kubenswrapper[4967]: I1008 00:19:25.152800 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b8900b3-7e99-48c8-a5a9-52a6edd4f7e1" containerName="registry-server" Oct 08 00:19:25 crc kubenswrapper[4967]: I1008 00:19:25.154375 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nns67" Oct 08 00:19:25 crc kubenswrapper[4967]: I1008 00:19:25.155067 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nns67"] Oct 08 00:19:25 crc kubenswrapper[4967]: I1008 00:19:25.176518 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0968c64a-5228-4e40-91e4-94a0b8ebc230-catalog-content\") pod \"redhat-operators-nns67\" (UID: \"0968c64a-5228-4e40-91e4-94a0b8ebc230\") " pod="openshift-marketplace/redhat-operators-nns67" Oct 08 00:19:25 crc kubenswrapper[4967]: I1008 00:19:25.176644 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0968c64a-5228-4e40-91e4-94a0b8ebc230-utilities\") pod \"redhat-operators-nns67\" (UID: \"0968c64a-5228-4e40-91e4-94a0b8ebc230\") " pod="openshift-marketplace/redhat-operators-nns67" Oct 08 00:19:25 crc kubenswrapper[4967]: I1008 00:19:25.176715 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rlb7s\" (UniqueName: \"kubernetes.io/projected/0968c64a-5228-4e40-91e4-94a0b8ebc230-kube-api-access-rlb7s\") pod \"redhat-operators-nns67\" (UID: \"0968c64a-5228-4e40-91e4-94a0b8ebc230\") " pod="openshift-marketplace/redhat-operators-nns67" Oct 08 00:19:25 crc kubenswrapper[4967]: I1008 00:19:25.277437 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0968c64a-5228-4e40-91e4-94a0b8ebc230-utilities\") pod \"redhat-operators-nns67\" (UID: \"0968c64a-5228-4e40-91e4-94a0b8ebc230\") " pod="openshift-marketplace/redhat-operators-nns67" Oct 08 00:19:25 crc kubenswrapper[4967]: I1008 00:19:25.277494 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rlb7s\" (UniqueName: \"kubernetes.io/projected/0968c64a-5228-4e40-91e4-94a0b8ebc230-kube-api-access-rlb7s\") pod \"redhat-operators-nns67\" (UID: \"0968c64a-5228-4e40-91e4-94a0b8ebc230\") " pod="openshift-marketplace/redhat-operators-nns67" Oct 08 00:19:25 crc kubenswrapper[4967]: I1008 00:19:25.277526 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0968c64a-5228-4e40-91e4-94a0b8ebc230-catalog-content\") pod \"redhat-operators-nns67\" (UID: \"0968c64a-5228-4e40-91e4-94a0b8ebc230\") " pod="openshift-marketplace/redhat-operators-nns67" Oct 08 00:19:25 crc kubenswrapper[4967]: I1008 00:19:25.278040 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0968c64a-5228-4e40-91e4-94a0b8ebc230-catalog-content\") pod \"redhat-operators-nns67\" (UID: \"0968c64a-5228-4e40-91e4-94a0b8ebc230\") " pod="openshift-marketplace/redhat-operators-nns67" Oct 08 00:19:25 crc kubenswrapper[4967]: I1008 00:19:25.278040 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0968c64a-5228-4e40-91e4-94a0b8ebc230-utilities\") pod \"redhat-operators-nns67\" (UID: \"0968c64a-5228-4e40-91e4-94a0b8ebc230\") " pod="openshift-marketplace/redhat-operators-nns67" Oct 08 00:19:25 crc kubenswrapper[4967]: I1008 00:19:25.298385 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rlb7s\" (UniqueName: \"kubernetes.io/projected/0968c64a-5228-4e40-91e4-94a0b8ebc230-kube-api-access-rlb7s\") pod \"redhat-operators-nns67\" (UID: \"0968c64a-5228-4e40-91e4-94a0b8ebc230\") " pod="openshift-marketplace/redhat-operators-nns67" Oct 08 00:19:25 crc kubenswrapper[4967]: I1008 00:19:25.509740 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nns67" Oct 08 00:19:25 crc kubenswrapper[4967]: I1008 00:19:25.914958 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nns67"] Oct 08 00:19:26 crc kubenswrapper[4967]: I1008 00:19:26.084112 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nns67" event={"ID":"0968c64a-5228-4e40-91e4-94a0b8ebc230","Type":"ContainerStarted","Data":"5d8d0365f830b5abd022a38e4d373904dabd91d12b1d9dbc5ac3659fce77895b"} Oct 08 00:19:27 crc kubenswrapper[4967]: I1008 00:19:27.092487 4967 generic.go:334] "Generic (PLEG): container finished" podID="0968c64a-5228-4e40-91e4-94a0b8ebc230" containerID="628385905ccc74c5b95a6152c2ee40300f417f83678db982d162710b3d8de3ce" exitCode=0 Oct 08 00:19:27 crc kubenswrapper[4967]: I1008 00:19:27.092610 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nns67" event={"ID":"0968c64a-5228-4e40-91e4-94a0b8ebc230","Type":"ContainerDied","Data":"628385905ccc74c5b95a6152c2ee40300f417f83678db982d162710b3d8de3ce"} Oct 08 00:19:28 crc kubenswrapper[4967]: I1008 00:19:28.107648 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nns67" event={"ID":"0968c64a-5228-4e40-91e4-94a0b8ebc230","Type":"ContainerStarted","Data":"9137c9e9916bdeda3bbc6914ee9cf66d2bb9efc388880ff19b8b71e2c47aa666"} Oct 08 00:19:29 crc kubenswrapper[4967]: I1008 00:19:29.116617 4967 generic.go:334] "Generic (PLEG): container finished" podID="0968c64a-5228-4e40-91e4-94a0b8ebc230" containerID="9137c9e9916bdeda3bbc6914ee9cf66d2bb9efc388880ff19b8b71e2c47aa666" exitCode=0 Oct 08 00:19:29 crc kubenswrapper[4967]: I1008 00:19:29.116726 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nns67" event={"ID":"0968c64a-5228-4e40-91e4-94a0b8ebc230","Type":"ContainerDied","Data":"9137c9e9916bdeda3bbc6914ee9cf66d2bb9efc388880ff19b8b71e2c47aa666"} Oct 08 00:19:30 crc kubenswrapper[4967]: I1008 00:19:30.132154 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nns67" event={"ID":"0968c64a-5228-4e40-91e4-94a0b8ebc230","Type":"ContainerStarted","Data":"00529d21e383b2b66e18ead2ec112b8c9df5a9a68887d6c6deacf7627bc56f51"} Oct 08 00:19:30 crc kubenswrapper[4967]: I1008 00:19:30.151950 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-nns67" podStartSLOduration=2.7390347569999998 podStartE2EDuration="5.151931985s" podCreationTimestamp="2025-10-08 00:19:25 +0000 UTC" firstStartedPulling="2025-10-08 00:19:27.094240639 +0000 UTC m=+765.744775328" lastFinishedPulling="2025-10-08 00:19:29.507137897 +0000 UTC m=+768.157672556" observedRunningTime="2025-10-08 00:19:30.147953276 +0000 UTC m=+768.798487925" watchObservedRunningTime="2025-10-08 00:19:30.151931985 +0000 UTC m=+768.802466644" Oct 08 00:19:35 crc kubenswrapper[4967]: I1008 00:19:35.510041 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-nns67" Oct 08 00:19:35 crc kubenswrapper[4967]: I1008 00:19:35.510354 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-nns67" Oct 08 00:19:35 crc kubenswrapper[4967]: I1008 00:19:35.570240 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-nns67" Oct 08 00:19:36 crc kubenswrapper[4967]: I1008 00:19:36.205880 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-nns67" Oct 08 00:19:36 crc kubenswrapper[4967]: I1008 00:19:36.250925 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nns67"] Oct 08 00:19:36 crc kubenswrapper[4967]: I1008 00:19:36.920156 4967 patch_prober.go:28] interesting pod/machine-config-daemon-xvdtp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 00:19:36 crc kubenswrapper[4967]: I1008 00:19:36.920219 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" podUID="9461162a-c27d-465e-9ca7-2d3936877cd8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 00:19:38 crc kubenswrapper[4967]: I1008 00:19:38.178949 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-nns67" podUID="0968c64a-5228-4e40-91e4-94a0b8ebc230" containerName="registry-server" containerID="cri-o://00529d21e383b2b66e18ead2ec112b8c9df5a9a68887d6c6deacf7627bc56f51" gracePeriod=2 Oct 08 00:19:39 crc kubenswrapper[4967]: I1008 00:19:39.075755 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nns67" Oct 08 00:19:39 crc kubenswrapper[4967]: I1008 00:19:39.185517 4967 generic.go:334] "Generic (PLEG): container finished" podID="0968c64a-5228-4e40-91e4-94a0b8ebc230" containerID="00529d21e383b2b66e18ead2ec112b8c9df5a9a68887d6c6deacf7627bc56f51" exitCode=0 Oct 08 00:19:39 crc kubenswrapper[4967]: I1008 00:19:39.185552 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nns67" event={"ID":"0968c64a-5228-4e40-91e4-94a0b8ebc230","Type":"ContainerDied","Data":"00529d21e383b2b66e18ead2ec112b8c9df5a9a68887d6c6deacf7627bc56f51"} Oct 08 00:19:39 crc kubenswrapper[4967]: I1008 00:19:39.185576 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nns67" event={"ID":"0968c64a-5228-4e40-91e4-94a0b8ebc230","Type":"ContainerDied","Data":"5d8d0365f830b5abd022a38e4d373904dabd91d12b1d9dbc5ac3659fce77895b"} Oct 08 00:19:39 crc kubenswrapper[4967]: I1008 00:19:39.185592 4967 scope.go:117] "RemoveContainer" containerID="00529d21e383b2b66e18ead2ec112b8c9df5a9a68887d6c6deacf7627bc56f51" Oct 08 00:19:39 crc kubenswrapper[4967]: I1008 00:19:39.185704 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nns67" Oct 08 00:19:39 crc kubenswrapper[4967]: I1008 00:19:39.194556 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rlb7s\" (UniqueName: \"kubernetes.io/projected/0968c64a-5228-4e40-91e4-94a0b8ebc230-kube-api-access-rlb7s\") pod \"0968c64a-5228-4e40-91e4-94a0b8ebc230\" (UID: \"0968c64a-5228-4e40-91e4-94a0b8ebc230\") " Oct 08 00:19:39 crc kubenswrapper[4967]: I1008 00:19:39.194650 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0968c64a-5228-4e40-91e4-94a0b8ebc230-utilities\") pod \"0968c64a-5228-4e40-91e4-94a0b8ebc230\" (UID: \"0968c64a-5228-4e40-91e4-94a0b8ebc230\") " Oct 08 00:19:39 crc kubenswrapper[4967]: I1008 00:19:39.194841 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0968c64a-5228-4e40-91e4-94a0b8ebc230-catalog-content\") pod \"0968c64a-5228-4e40-91e4-94a0b8ebc230\" (UID: \"0968c64a-5228-4e40-91e4-94a0b8ebc230\") " Oct 08 00:19:39 crc kubenswrapper[4967]: I1008 00:19:39.195915 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0968c64a-5228-4e40-91e4-94a0b8ebc230-utilities" (OuterVolumeSpecName: "utilities") pod "0968c64a-5228-4e40-91e4-94a0b8ebc230" (UID: "0968c64a-5228-4e40-91e4-94a0b8ebc230"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:19:39 crc kubenswrapper[4967]: I1008 00:19:39.218031 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0968c64a-5228-4e40-91e4-94a0b8ebc230-kube-api-access-rlb7s" (OuterVolumeSpecName: "kube-api-access-rlb7s") pod "0968c64a-5228-4e40-91e4-94a0b8ebc230" (UID: "0968c64a-5228-4e40-91e4-94a0b8ebc230"). InnerVolumeSpecName "kube-api-access-rlb7s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:19:39 crc kubenswrapper[4967]: I1008 00:19:39.223331 4967 scope.go:117] "RemoveContainer" containerID="9137c9e9916bdeda3bbc6914ee9cf66d2bb9efc388880ff19b8b71e2c47aa666" Oct 08 00:19:39 crc kubenswrapper[4967]: I1008 00:19:39.254348 4967 scope.go:117] "RemoveContainer" containerID="628385905ccc74c5b95a6152c2ee40300f417f83678db982d162710b3d8de3ce" Oct 08 00:19:39 crc kubenswrapper[4967]: I1008 00:19:39.284218 4967 scope.go:117] "RemoveContainer" containerID="00529d21e383b2b66e18ead2ec112b8c9df5a9a68887d6c6deacf7627bc56f51" Oct 08 00:19:39 crc kubenswrapper[4967]: E1008 00:19:39.287642 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"00529d21e383b2b66e18ead2ec112b8c9df5a9a68887d6c6deacf7627bc56f51\": container with ID starting with 00529d21e383b2b66e18ead2ec112b8c9df5a9a68887d6c6deacf7627bc56f51 not found: ID does not exist" containerID="00529d21e383b2b66e18ead2ec112b8c9df5a9a68887d6c6deacf7627bc56f51" Oct 08 00:19:39 crc kubenswrapper[4967]: I1008 00:19:39.287683 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00529d21e383b2b66e18ead2ec112b8c9df5a9a68887d6c6deacf7627bc56f51"} err="failed to get container status \"00529d21e383b2b66e18ead2ec112b8c9df5a9a68887d6c6deacf7627bc56f51\": rpc error: code = NotFound desc = could not find container \"00529d21e383b2b66e18ead2ec112b8c9df5a9a68887d6c6deacf7627bc56f51\": container with ID starting with 00529d21e383b2b66e18ead2ec112b8c9df5a9a68887d6c6deacf7627bc56f51 not found: ID does not exist" Oct 08 00:19:39 crc kubenswrapper[4967]: I1008 00:19:39.287728 4967 scope.go:117] "RemoveContainer" containerID="9137c9e9916bdeda3bbc6914ee9cf66d2bb9efc388880ff19b8b71e2c47aa666" Oct 08 00:19:39 crc kubenswrapper[4967]: E1008 00:19:39.287989 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9137c9e9916bdeda3bbc6914ee9cf66d2bb9efc388880ff19b8b71e2c47aa666\": container with ID starting with 9137c9e9916bdeda3bbc6914ee9cf66d2bb9efc388880ff19b8b71e2c47aa666 not found: ID does not exist" containerID="9137c9e9916bdeda3bbc6914ee9cf66d2bb9efc388880ff19b8b71e2c47aa666" Oct 08 00:19:39 crc kubenswrapper[4967]: I1008 00:19:39.288009 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9137c9e9916bdeda3bbc6914ee9cf66d2bb9efc388880ff19b8b71e2c47aa666"} err="failed to get container status \"9137c9e9916bdeda3bbc6914ee9cf66d2bb9efc388880ff19b8b71e2c47aa666\": rpc error: code = NotFound desc = could not find container \"9137c9e9916bdeda3bbc6914ee9cf66d2bb9efc388880ff19b8b71e2c47aa666\": container with ID starting with 9137c9e9916bdeda3bbc6914ee9cf66d2bb9efc388880ff19b8b71e2c47aa666 not found: ID does not exist" Oct 08 00:19:39 crc kubenswrapper[4967]: I1008 00:19:39.288025 4967 scope.go:117] "RemoveContainer" containerID="628385905ccc74c5b95a6152c2ee40300f417f83678db982d162710b3d8de3ce" Oct 08 00:19:39 crc kubenswrapper[4967]: E1008 00:19:39.288284 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"628385905ccc74c5b95a6152c2ee40300f417f83678db982d162710b3d8de3ce\": container with ID starting with 628385905ccc74c5b95a6152c2ee40300f417f83678db982d162710b3d8de3ce not found: ID does not exist" containerID="628385905ccc74c5b95a6152c2ee40300f417f83678db982d162710b3d8de3ce" Oct 08 00:19:39 crc kubenswrapper[4967]: I1008 00:19:39.288305 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"628385905ccc74c5b95a6152c2ee40300f417f83678db982d162710b3d8de3ce"} err="failed to get container status \"628385905ccc74c5b95a6152c2ee40300f417f83678db982d162710b3d8de3ce\": rpc error: code = NotFound desc = could not find container \"628385905ccc74c5b95a6152c2ee40300f417f83678db982d162710b3d8de3ce\": container with ID starting with 628385905ccc74c5b95a6152c2ee40300f417f83678db982d162710b3d8de3ce not found: ID does not exist" Oct 08 00:19:39 crc kubenswrapper[4967]: I1008 00:19:39.296360 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rlb7s\" (UniqueName: \"kubernetes.io/projected/0968c64a-5228-4e40-91e4-94a0b8ebc230-kube-api-access-rlb7s\") on node \"crc\" DevicePath \"\"" Oct 08 00:19:39 crc kubenswrapper[4967]: I1008 00:19:39.296394 4967 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0968c64a-5228-4e40-91e4-94a0b8ebc230-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 00:19:39 crc kubenswrapper[4967]: I1008 00:19:39.863319 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0968c64a-5228-4e40-91e4-94a0b8ebc230-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0968c64a-5228-4e40-91e4-94a0b8ebc230" (UID: "0968c64a-5228-4e40-91e4-94a0b8ebc230"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:19:39 crc kubenswrapper[4967]: I1008 00:19:39.905004 4967 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0968c64a-5228-4e40-91e4-94a0b8ebc230-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 00:19:40 crc kubenswrapper[4967]: I1008 00:19:40.130906 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nns67"] Oct 08 00:19:40 crc kubenswrapper[4967]: I1008 00:19:40.135106 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-nns67"] Oct 08 00:19:42 crc kubenswrapper[4967]: I1008 00:19:42.125645 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0968c64a-5228-4e40-91e4-94a0b8ebc230" path="/var/lib/kubelet/pods/0968c64a-5228-4e40-91e4-94a0b8ebc230/volumes" Oct 08 00:20:06 crc kubenswrapper[4967]: I1008 00:20:06.919735 4967 patch_prober.go:28] interesting pod/machine-config-daemon-xvdtp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 00:20:06 crc kubenswrapper[4967]: I1008 00:20:06.920324 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" podUID="9461162a-c27d-465e-9ca7-2d3936877cd8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 00:20:13 crc kubenswrapper[4967]: I1008 00:20:13.412382 4967 generic.go:334] "Generic (PLEG): container finished" podID="76078656-a41c-4e34-ad5d-ca11ebaa3ff3" containerID="0a43c71e4932386d7fed98c8560524c1d4b0a845f0edd0e8b48a515f27036226" exitCode=0 Oct 08 00:20:13 crc kubenswrapper[4967]: I1008 00:20:13.412486 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-2-build" event={"ID":"76078656-a41c-4e34-ad5d-ca11ebaa3ff3","Type":"ContainerDied","Data":"0a43c71e4932386d7fed98c8560524c1d4b0a845f0edd0e8b48a515f27036226"} Oct 08 00:20:14 crc kubenswrapper[4967]: I1008 00:20:14.676881 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-2-build" Oct 08 00:20:14 crc kubenswrapper[4967]: I1008 00:20:14.817710 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/76078656-a41c-4e34-ad5d-ca11ebaa3ff3-build-blob-cache\") pod \"76078656-a41c-4e34-ad5d-ca11ebaa3ff3\" (UID: \"76078656-a41c-4e34-ad5d-ca11ebaa3ff3\") " Oct 08 00:20:14 crc kubenswrapper[4967]: I1008 00:20:14.817760 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bqgnl\" (UniqueName: \"kubernetes.io/projected/76078656-a41c-4e34-ad5d-ca11ebaa3ff3-kube-api-access-bqgnl\") pod \"76078656-a41c-4e34-ad5d-ca11ebaa3ff3\" (UID: \"76078656-a41c-4e34-ad5d-ca11ebaa3ff3\") " Oct 08 00:20:14 crc kubenswrapper[4967]: I1008 00:20:14.817826 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/76078656-a41c-4e34-ad5d-ca11ebaa3ff3-buildcachedir\") pod \"76078656-a41c-4e34-ad5d-ca11ebaa3ff3\" (UID: \"76078656-a41c-4e34-ad5d-ca11ebaa3ff3\") " Oct 08 00:20:14 crc kubenswrapper[4967]: I1008 00:20:14.817860 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-r7snr-pull\" (UniqueName: \"kubernetes.io/secret/76078656-a41c-4e34-ad5d-ca11ebaa3ff3-builder-dockercfg-r7snr-pull\") pod \"76078656-a41c-4e34-ad5d-ca11ebaa3ff3\" (UID: \"76078656-a41c-4e34-ad5d-ca11ebaa3ff3\") " Oct 08 00:20:14 crc kubenswrapper[4967]: I1008 00:20:14.817933 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/76078656-a41c-4e34-ad5d-ca11ebaa3ff3-node-pullsecrets\") pod \"76078656-a41c-4e34-ad5d-ca11ebaa3ff3\" (UID: \"76078656-a41c-4e34-ad5d-ca11ebaa3ff3\") " Oct 08 00:20:14 crc kubenswrapper[4967]: I1008 00:20:14.817908 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/76078656-a41c-4e34-ad5d-ca11ebaa3ff3-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "76078656-a41c-4e34-ad5d-ca11ebaa3ff3" (UID: "76078656-a41c-4e34-ad5d-ca11ebaa3ff3"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 00:20:14 crc kubenswrapper[4967]: I1008 00:20:14.817998 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/76078656-a41c-4e34-ad5d-ca11ebaa3ff3-container-storage-root\") pod \"76078656-a41c-4e34-ad5d-ca11ebaa3ff3\" (UID: \"76078656-a41c-4e34-ad5d-ca11ebaa3ff3\") " Oct 08 00:20:14 crc kubenswrapper[4967]: I1008 00:20:14.818039 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/76078656-a41c-4e34-ad5d-ca11ebaa3ff3-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "76078656-a41c-4e34-ad5d-ca11ebaa3ff3" (UID: "76078656-a41c-4e34-ad5d-ca11ebaa3ff3"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 00:20:14 crc kubenswrapper[4967]: I1008 00:20:14.820817 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/76078656-a41c-4e34-ad5d-ca11ebaa3ff3-buildworkdir\") pod \"76078656-a41c-4e34-ad5d-ca11ebaa3ff3\" (UID: \"76078656-a41c-4e34-ad5d-ca11ebaa3ff3\") " Oct 08 00:20:14 crc kubenswrapper[4967]: I1008 00:20:14.821182 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/76078656-a41c-4e34-ad5d-ca11ebaa3ff3-build-ca-bundles\") pod \"76078656-a41c-4e34-ad5d-ca11ebaa3ff3\" (UID: \"76078656-a41c-4e34-ad5d-ca11ebaa3ff3\") " Oct 08 00:20:14 crc kubenswrapper[4967]: I1008 00:20:14.821294 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-r7snr-push\" (UniqueName: \"kubernetes.io/secret/76078656-a41c-4e34-ad5d-ca11ebaa3ff3-builder-dockercfg-r7snr-push\") pod \"76078656-a41c-4e34-ad5d-ca11ebaa3ff3\" (UID: \"76078656-a41c-4e34-ad5d-ca11ebaa3ff3\") " Oct 08 00:20:14 crc kubenswrapper[4967]: I1008 00:20:14.821366 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/76078656-a41c-4e34-ad5d-ca11ebaa3ff3-container-storage-run\") pod \"76078656-a41c-4e34-ad5d-ca11ebaa3ff3\" (UID: \"76078656-a41c-4e34-ad5d-ca11ebaa3ff3\") " Oct 08 00:20:14 crc kubenswrapper[4967]: I1008 00:20:14.821430 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/76078656-a41c-4e34-ad5d-ca11ebaa3ff3-build-system-configs\") pod \"76078656-a41c-4e34-ad5d-ca11ebaa3ff3\" (UID: \"76078656-a41c-4e34-ad5d-ca11ebaa3ff3\") " Oct 08 00:20:14 crc kubenswrapper[4967]: I1008 00:20:14.821526 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/76078656-a41c-4e34-ad5d-ca11ebaa3ff3-build-proxy-ca-bundles\") pod \"76078656-a41c-4e34-ad5d-ca11ebaa3ff3\" (UID: \"76078656-a41c-4e34-ad5d-ca11ebaa3ff3\") " Oct 08 00:20:14 crc kubenswrapper[4967]: I1008 00:20:14.822234 4967 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/76078656-a41c-4e34-ad5d-ca11ebaa3ff3-buildcachedir\") on node \"crc\" DevicePath \"\"" Oct 08 00:20:14 crc kubenswrapper[4967]: I1008 00:20:14.822320 4967 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/76078656-a41c-4e34-ad5d-ca11ebaa3ff3-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Oct 08 00:20:14 crc kubenswrapper[4967]: I1008 00:20:14.825147 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/76078656-a41c-4e34-ad5d-ca11ebaa3ff3-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "76078656-a41c-4e34-ad5d-ca11ebaa3ff3" (UID: "76078656-a41c-4e34-ad5d-ca11ebaa3ff3"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:20:14 crc kubenswrapper[4967]: I1008 00:20:14.826001 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76078656-a41c-4e34-ad5d-ca11ebaa3ff3-kube-api-access-bqgnl" (OuterVolumeSpecName: "kube-api-access-bqgnl") pod "76078656-a41c-4e34-ad5d-ca11ebaa3ff3" (UID: "76078656-a41c-4e34-ad5d-ca11ebaa3ff3"). InnerVolumeSpecName "kube-api-access-bqgnl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:20:14 crc kubenswrapper[4967]: I1008 00:20:14.826095 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76078656-a41c-4e34-ad5d-ca11ebaa3ff3-builder-dockercfg-r7snr-pull" (OuterVolumeSpecName: "builder-dockercfg-r7snr-pull") pod "76078656-a41c-4e34-ad5d-ca11ebaa3ff3" (UID: "76078656-a41c-4e34-ad5d-ca11ebaa3ff3"). InnerVolumeSpecName "builder-dockercfg-r7snr-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:20:14 crc kubenswrapper[4967]: I1008 00:20:14.827870 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76078656-a41c-4e34-ad5d-ca11ebaa3ff3-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "76078656-a41c-4e34-ad5d-ca11ebaa3ff3" (UID: "76078656-a41c-4e34-ad5d-ca11ebaa3ff3"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:20:14 crc kubenswrapper[4967]: I1008 00:20:14.829239 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/76078656-a41c-4e34-ad5d-ca11ebaa3ff3-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "76078656-a41c-4e34-ad5d-ca11ebaa3ff3" (UID: "76078656-a41c-4e34-ad5d-ca11ebaa3ff3"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:20:14 crc kubenswrapper[4967]: I1008 00:20:14.830401 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/76078656-a41c-4e34-ad5d-ca11ebaa3ff3-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "76078656-a41c-4e34-ad5d-ca11ebaa3ff3" (UID: "76078656-a41c-4e34-ad5d-ca11ebaa3ff3"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:20:14 crc kubenswrapper[4967]: I1008 00:20:14.838936 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76078656-a41c-4e34-ad5d-ca11ebaa3ff3-builder-dockercfg-r7snr-push" (OuterVolumeSpecName: "builder-dockercfg-r7snr-push") pod "76078656-a41c-4e34-ad5d-ca11ebaa3ff3" (UID: "76078656-a41c-4e34-ad5d-ca11ebaa3ff3"). InnerVolumeSpecName "builder-dockercfg-r7snr-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:20:14 crc kubenswrapper[4967]: I1008 00:20:14.865758 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76078656-a41c-4e34-ad5d-ca11ebaa3ff3-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "76078656-a41c-4e34-ad5d-ca11ebaa3ff3" (UID: "76078656-a41c-4e34-ad5d-ca11ebaa3ff3"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:20:14 crc kubenswrapper[4967]: I1008 00:20:14.924093 4967 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/76078656-a41c-4e34-ad5d-ca11ebaa3ff3-build-system-configs\") on node \"crc\" DevicePath \"\"" Oct 08 00:20:14 crc kubenswrapper[4967]: I1008 00:20:14.924127 4967 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-r7snr-push\" (UniqueName: \"kubernetes.io/secret/76078656-a41c-4e34-ad5d-ca11ebaa3ff3-builder-dockercfg-r7snr-push\") on node \"crc\" DevicePath \"\"" Oct 08 00:20:14 crc kubenswrapper[4967]: I1008 00:20:14.924141 4967 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/76078656-a41c-4e34-ad5d-ca11ebaa3ff3-container-storage-run\") on node \"crc\" DevicePath \"\"" Oct 08 00:20:14 crc kubenswrapper[4967]: I1008 00:20:14.924152 4967 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/76078656-a41c-4e34-ad5d-ca11ebaa3ff3-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 08 00:20:14 crc kubenswrapper[4967]: I1008 00:20:14.924164 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bqgnl\" (UniqueName: \"kubernetes.io/projected/76078656-a41c-4e34-ad5d-ca11ebaa3ff3-kube-api-access-bqgnl\") on node \"crc\" DevicePath \"\"" Oct 08 00:20:14 crc kubenswrapper[4967]: I1008 00:20:14.924175 4967 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-r7snr-pull\" (UniqueName: \"kubernetes.io/secret/76078656-a41c-4e34-ad5d-ca11ebaa3ff3-builder-dockercfg-r7snr-pull\") on node \"crc\" DevicePath \"\"" Oct 08 00:20:14 crc kubenswrapper[4967]: I1008 00:20:14.924187 4967 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/76078656-a41c-4e34-ad5d-ca11ebaa3ff3-buildworkdir\") on node \"crc\" DevicePath \"\"" Oct 08 00:20:14 crc kubenswrapper[4967]: I1008 00:20:14.924200 4967 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/76078656-a41c-4e34-ad5d-ca11ebaa3ff3-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 08 00:20:15 crc kubenswrapper[4967]: I1008 00:20:15.018377 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76078656-a41c-4e34-ad5d-ca11ebaa3ff3-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "76078656-a41c-4e34-ad5d-ca11ebaa3ff3" (UID: "76078656-a41c-4e34-ad5d-ca11ebaa3ff3"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:20:15 crc kubenswrapper[4967]: I1008 00:20:15.025900 4967 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/76078656-a41c-4e34-ad5d-ca11ebaa3ff3-build-blob-cache\") on node \"crc\" DevicePath \"\"" Oct 08 00:20:15 crc kubenswrapper[4967]: I1008 00:20:15.427365 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-2-build" event={"ID":"76078656-a41c-4e34-ad5d-ca11ebaa3ff3","Type":"ContainerDied","Data":"ea9a167c36c73916cee3e8f90dca201b4778ac87941ae7f63d63f3821fd84fe2"} Oct 08 00:20:15 crc kubenswrapper[4967]: I1008 00:20:15.427420 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-2-build" Oct 08 00:20:15 crc kubenswrapper[4967]: I1008 00:20:15.427426 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ea9a167c36c73916cee3e8f90dca201b4778ac87941ae7f63d63f3821fd84fe2" Oct 08 00:20:17 crc kubenswrapper[4967]: I1008 00:20:17.082432 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76078656-a41c-4e34-ad5d-ca11ebaa3ff3-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "76078656-a41c-4e34-ad5d-ca11ebaa3ff3" (UID: "76078656-a41c-4e34-ad5d-ca11ebaa3ff3"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:20:17 crc kubenswrapper[4967]: I1008 00:20:17.157614 4967 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/76078656-a41c-4e34-ad5d-ca11ebaa3ff3-container-storage-root\") on node \"crc\" DevicePath \"\"" Oct 08 00:20:19 crc kubenswrapper[4967]: I1008 00:20:19.502989 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/smart-gateway-operator-1-build"] Oct 08 00:20:19 crc kubenswrapper[4967]: E1008 00:20:19.503290 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76078656-a41c-4e34-ad5d-ca11ebaa3ff3" containerName="docker-build" Oct 08 00:20:19 crc kubenswrapper[4967]: I1008 00:20:19.506171 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="76078656-a41c-4e34-ad5d-ca11ebaa3ff3" containerName="docker-build" Oct 08 00:20:19 crc kubenswrapper[4967]: E1008 00:20:19.506201 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76078656-a41c-4e34-ad5d-ca11ebaa3ff3" containerName="git-clone" Oct 08 00:20:19 crc kubenswrapper[4967]: I1008 00:20:19.506212 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="76078656-a41c-4e34-ad5d-ca11ebaa3ff3" containerName="git-clone" Oct 08 00:20:19 crc kubenswrapper[4967]: E1008 00:20:19.506233 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0968c64a-5228-4e40-91e4-94a0b8ebc230" containerName="extract-utilities" Oct 08 00:20:19 crc kubenswrapper[4967]: I1008 00:20:19.506246 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="0968c64a-5228-4e40-91e4-94a0b8ebc230" containerName="extract-utilities" Oct 08 00:20:19 crc kubenswrapper[4967]: E1008 00:20:19.506269 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76078656-a41c-4e34-ad5d-ca11ebaa3ff3" containerName="manage-dockerfile" Oct 08 00:20:19 crc kubenswrapper[4967]: I1008 00:20:19.506279 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="76078656-a41c-4e34-ad5d-ca11ebaa3ff3" containerName="manage-dockerfile" Oct 08 00:20:19 crc kubenswrapper[4967]: E1008 00:20:19.506297 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0968c64a-5228-4e40-91e4-94a0b8ebc230" containerName="extract-content" Oct 08 00:20:19 crc kubenswrapper[4967]: I1008 00:20:19.506307 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="0968c64a-5228-4e40-91e4-94a0b8ebc230" containerName="extract-content" Oct 08 00:20:19 crc kubenswrapper[4967]: E1008 00:20:19.506323 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0968c64a-5228-4e40-91e4-94a0b8ebc230" containerName="registry-server" Oct 08 00:20:19 crc kubenswrapper[4967]: I1008 00:20:19.506333 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="0968c64a-5228-4e40-91e4-94a0b8ebc230" containerName="registry-server" Oct 08 00:20:19 crc kubenswrapper[4967]: I1008 00:20:19.506497 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="0968c64a-5228-4e40-91e4-94a0b8ebc230" containerName="registry-server" Oct 08 00:20:19 crc kubenswrapper[4967]: I1008 00:20:19.506520 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="76078656-a41c-4e34-ad5d-ca11ebaa3ff3" containerName="docker-build" Oct 08 00:20:19 crc kubenswrapper[4967]: I1008 00:20:19.507365 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-1-build" Oct 08 00:20:19 crc kubenswrapper[4967]: I1008 00:20:19.509402 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"smart-gateway-operator-1-sys-config" Oct 08 00:20:19 crc kubenswrapper[4967]: I1008 00:20:19.509524 4967 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-r7snr" Oct 08 00:20:19 crc kubenswrapper[4967]: I1008 00:20:19.511301 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"smart-gateway-operator-1-ca" Oct 08 00:20:19 crc kubenswrapper[4967]: I1008 00:20:19.511554 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"smart-gateway-operator-1-global-ca" Oct 08 00:20:19 crc kubenswrapper[4967]: I1008 00:20:19.526094 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/smart-gateway-operator-1-build"] Oct 08 00:20:19 crc kubenswrapper[4967]: I1008 00:20:19.597868 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/f47a4120-aa3b-400b-aa4f-42b3262efd73-container-storage-run\") pod \"smart-gateway-operator-1-build\" (UID: \"f47a4120-aa3b-400b-aa4f-42b3262efd73\") " pod="service-telemetry/smart-gateway-operator-1-build" Oct 08 00:20:19 crc kubenswrapper[4967]: I1008 00:20:19.597941 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-r7snr-pull\" (UniqueName: \"kubernetes.io/secret/f47a4120-aa3b-400b-aa4f-42b3262efd73-builder-dockercfg-r7snr-pull\") pod \"smart-gateway-operator-1-build\" (UID: \"f47a4120-aa3b-400b-aa4f-42b3262efd73\") " pod="service-telemetry/smart-gateway-operator-1-build" Oct 08 00:20:19 crc kubenswrapper[4967]: I1008 00:20:19.598069 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f47a4120-aa3b-400b-aa4f-42b3262efd73-build-ca-bundles\") pod \"smart-gateway-operator-1-build\" (UID: \"f47a4120-aa3b-400b-aa4f-42b3262efd73\") " pod="service-telemetry/smart-gateway-operator-1-build" Oct 08 00:20:19 crc kubenswrapper[4967]: I1008 00:20:19.598196 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/f47a4120-aa3b-400b-aa4f-42b3262efd73-buildcachedir\") pod \"smart-gateway-operator-1-build\" (UID: \"f47a4120-aa3b-400b-aa4f-42b3262efd73\") " pod="service-telemetry/smart-gateway-operator-1-build" Oct 08 00:20:19 crc kubenswrapper[4967]: I1008 00:20:19.598251 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/f47a4120-aa3b-400b-aa4f-42b3262efd73-buildworkdir\") pod \"smart-gateway-operator-1-build\" (UID: \"f47a4120-aa3b-400b-aa4f-42b3262efd73\") " pod="service-telemetry/smart-gateway-operator-1-build" Oct 08 00:20:19 crc kubenswrapper[4967]: I1008 00:20:19.598332 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-r7snr-push\" (UniqueName: \"kubernetes.io/secret/f47a4120-aa3b-400b-aa4f-42b3262efd73-builder-dockercfg-r7snr-push\") pod \"smart-gateway-operator-1-build\" (UID: \"f47a4120-aa3b-400b-aa4f-42b3262efd73\") " pod="service-telemetry/smart-gateway-operator-1-build" Oct 08 00:20:19 crc kubenswrapper[4967]: I1008 00:20:19.598383 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j8rnd\" (UniqueName: \"kubernetes.io/projected/f47a4120-aa3b-400b-aa4f-42b3262efd73-kube-api-access-j8rnd\") pod \"smart-gateway-operator-1-build\" (UID: \"f47a4120-aa3b-400b-aa4f-42b3262efd73\") " pod="service-telemetry/smart-gateway-operator-1-build" Oct 08 00:20:19 crc kubenswrapper[4967]: I1008 00:20:19.598496 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f47a4120-aa3b-400b-aa4f-42b3262efd73-build-proxy-ca-bundles\") pod \"smart-gateway-operator-1-build\" (UID: \"f47a4120-aa3b-400b-aa4f-42b3262efd73\") " pod="service-telemetry/smart-gateway-operator-1-build" Oct 08 00:20:19 crc kubenswrapper[4967]: I1008 00:20:19.598551 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/f47a4120-aa3b-400b-aa4f-42b3262efd73-node-pullsecrets\") pod \"smart-gateway-operator-1-build\" (UID: \"f47a4120-aa3b-400b-aa4f-42b3262efd73\") " pod="service-telemetry/smart-gateway-operator-1-build" Oct 08 00:20:19 crc kubenswrapper[4967]: I1008 00:20:19.598632 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/f47a4120-aa3b-400b-aa4f-42b3262efd73-build-blob-cache\") pod \"smart-gateway-operator-1-build\" (UID: \"f47a4120-aa3b-400b-aa4f-42b3262efd73\") " pod="service-telemetry/smart-gateway-operator-1-build" Oct 08 00:20:19 crc kubenswrapper[4967]: I1008 00:20:19.598731 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/f47a4120-aa3b-400b-aa4f-42b3262efd73-build-system-configs\") pod \"smart-gateway-operator-1-build\" (UID: \"f47a4120-aa3b-400b-aa4f-42b3262efd73\") " pod="service-telemetry/smart-gateway-operator-1-build" Oct 08 00:20:19 crc kubenswrapper[4967]: I1008 00:20:19.598781 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/f47a4120-aa3b-400b-aa4f-42b3262efd73-container-storage-root\") pod \"smart-gateway-operator-1-build\" (UID: \"f47a4120-aa3b-400b-aa4f-42b3262efd73\") " pod="service-telemetry/smart-gateway-operator-1-build" Oct 08 00:20:19 crc kubenswrapper[4967]: I1008 00:20:19.700239 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/f47a4120-aa3b-400b-aa4f-42b3262efd73-buildcachedir\") pod \"smart-gateway-operator-1-build\" (UID: \"f47a4120-aa3b-400b-aa4f-42b3262efd73\") " pod="service-telemetry/smart-gateway-operator-1-build" Oct 08 00:20:19 crc kubenswrapper[4967]: I1008 00:20:19.700339 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/f47a4120-aa3b-400b-aa4f-42b3262efd73-buildcachedir\") pod \"smart-gateway-operator-1-build\" (UID: \"f47a4120-aa3b-400b-aa4f-42b3262efd73\") " pod="service-telemetry/smart-gateway-operator-1-build" Oct 08 00:20:19 crc kubenswrapper[4967]: I1008 00:20:19.700438 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/f47a4120-aa3b-400b-aa4f-42b3262efd73-buildworkdir\") pod \"smart-gateway-operator-1-build\" (UID: \"f47a4120-aa3b-400b-aa4f-42b3262efd73\") " pod="service-telemetry/smart-gateway-operator-1-build" Oct 08 00:20:19 crc kubenswrapper[4967]: I1008 00:20:19.701164 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/f47a4120-aa3b-400b-aa4f-42b3262efd73-buildworkdir\") pod \"smart-gateway-operator-1-build\" (UID: \"f47a4120-aa3b-400b-aa4f-42b3262efd73\") " pod="service-telemetry/smart-gateway-operator-1-build" Oct 08 00:20:19 crc kubenswrapper[4967]: I1008 00:20:19.700547 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-r7snr-push\" (UniqueName: \"kubernetes.io/secret/f47a4120-aa3b-400b-aa4f-42b3262efd73-builder-dockercfg-r7snr-push\") pod \"smart-gateway-operator-1-build\" (UID: \"f47a4120-aa3b-400b-aa4f-42b3262efd73\") " pod="service-telemetry/smart-gateway-operator-1-build" Oct 08 00:20:19 crc kubenswrapper[4967]: I1008 00:20:19.701296 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j8rnd\" (UniqueName: \"kubernetes.io/projected/f47a4120-aa3b-400b-aa4f-42b3262efd73-kube-api-access-j8rnd\") pod \"smart-gateway-operator-1-build\" (UID: \"f47a4120-aa3b-400b-aa4f-42b3262efd73\") " pod="service-telemetry/smart-gateway-operator-1-build" Oct 08 00:20:19 crc kubenswrapper[4967]: I1008 00:20:19.701942 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f47a4120-aa3b-400b-aa4f-42b3262efd73-build-proxy-ca-bundles\") pod \"smart-gateway-operator-1-build\" (UID: \"f47a4120-aa3b-400b-aa4f-42b3262efd73\") " pod="service-telemetry/smart-gateway-operator-1-build" Oct 08 00:20:19 crc kubenswrapper[4967]: I1008 00:20:19.702901 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/f47a4120-aa3b-400b-aa4f-42b3262efd73-node-pullsecrets\") pod \"smart-gateway-operator-1-build\" (UID: \"f47a4120-aa3b-400b-aa4f-42b3262efd73\") " pod="service-telemetry/smart-gateway-operator-1-build" Oct 08 00:20:19 crc kubenswrapper[4967]: I1008 00:20:19.702944 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f47a4120-aa3b-400b-aa4f-42b3262efd73-build-proxy-ca-bundles\") pod \"smart-gateway-operator-1-build\" (UID: \"f47a4120-aa3b-400b-aa4f-42b3262efd73\") " pod="service-telemetry/smart-gateway-operator-1-build" Oct 08 00:20:19 crc kubenswrapper[4967]: I1008 00:20:19.702975 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/f47a4120-aa3b-400b-aa4f-42b3262efd73-build-blob-cache\") pod \"smart-gateway-operator-1-build\" (UID: \"f47a4120-aa3b-400b-aa4f-42b3262efd73\") " pod="service-telemetry/smart-gateway-operator-1-build" Oct 08 00:20:19 crc kubenswrapper[4967]: I1008 00:20:19.703039 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/f47a4120-aa3b-400b-aa4f-42b3262efd73-build-system-configs\") pod \"smart-gateway-operator-1-build\" (UID: \"f47a4120-aa3b-400b-aa4f-42b3262efd73\") " pod="service-telemetry/smart-gateway-operator-1-build" Oct 08 00:20:19 crc kubenswrapper[4967]: I1008 00:20:19.703107 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/f47a4120-aa3b-400b-aa4f-42b3262efd73-container-storage-root\") pod \"smart-gateway-operator-1-build\" (UID: \"f47a4120-aa3b-400b-aa4f-42b3262efd73\") " pod="service-telemetry/smart-gateway-operator-1-build" Oct 08 00:20:19 crc kubenswrapper[4967]: I1008 00:20:19.703141 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/f47a4120-aa3b-400b-aa4f-42b3262efd73-node-pullsecrets\") pod \"smart-gateway-operator-1-build\" (UID: \"f47a4120-aa3b-400b-aa4f-42b3262efd73\") " pod="service-telemetry/smart-gateway-operator-1-build" Oct 08 00:20:19 crc kubenswrapper[4967]: I1008 00:20:19.703219 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/f47a4120-aa3b-400b-aa4f-42b3262efd73-container-storage-run\") pod \"smart-gateway-operator-1-build\" (UID: \"f47a4120-aa3b-400b-aa4f-42b3262efd73\") " pod="service-telemetry/smart-gateway-operator-1-build" Oct 08 00:20:19 crc kubenswrapper[4967]: I1008 00:20:19.703282 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-r7snr-pull\" (UniqueName: \"kubernetes.io/secret/f47a4120-aa3b-400b-aa4f-42b3262efd73-builder-dockercfg-r7snr-pull\") pod \"smart-gateway-operator-1-build\" (UID: \"f47a4120-aa3b-400b-aa4f-42b3262efd73\") " pod="service-telemetry/smart-gateway-operator-1-build" Oct 08 00:20:19 crc kubenswrapper[4967]: I1008 00:20:19.703388 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/f47a4120-aa3b-400b-aa4f-42b3262efd73-build-blob-cache\") pod \"smart-gateway-operator-1-build\" (UID: \"f47a4120-aa3b-400b-aa4f-42b3262efd73\") " pod="service-telemetry/smart-gateway-operator-1-build" Oct 08 00:20:19 crc kubenswrapper[4967]: I1008 00:20:19.703585 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f47a4120-aa3b-400b-aa4f-42b3262efd73-build-ca-bundles\") pod \"smart-gateway-operator-1-build\" (UID: \"f47a4120-aa3b-400b-aa4f-42b3262efd73\") " pod="service-telemetry/smart-gateway-operator-1-build" Oct 08 00:20:19 crc kubenswrapper[4967]: I1008 00:20:19.704009 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/f47a4120-aa3b-400b-aa4f-42b3262efd73-container-storage-root\") pod \"smart-gateway-operator-1-build\" (UID: \"f47a4120-aa3b-400b-aa4f-42b3262efd73\") " pod="service-telemetry/smart-gateway-operator-1-build" Oct 08 00:20:19 crc kubenswrapper[4967]: I1008 00:20:19.704226 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/f47a4120-aa3b-400b-aa4f-42b3262efd73-container-storage-run\") pod \"smart-gateway-operator-1-build\" (UID: \"f47a4120-aa3b-400b-aa4f-42b3262efd73\") " pod="service-telemetry/smart-gateway-operator-1-build" Oct 08 00:20:19 crc kubenswrapper[4967]: I1008 00:20:19.705105 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/f47a4120-aa3b-400b-aa4f-42b3262efd73-build-system-configs\") pod \"smart-gateway-operator-1-build\" (UID: \"f47a4120-aa3b-400b-aa4f-42b3262efd73\") " pod="service-telemetry/smart-gateway-operator-1-build" Oct 08 00:20:19 crc kubenswrapper[4967]: I1008 00:20:19.705605 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f47a4120-aa3b-400b-aa4f-42b3262efd73-build-ca-bundles\") pod \"smart-gateway-operator-1-build\" (UID: \"f47a4120-aa3b-400b-aa4f-42b3262efd73\") " pod="service-telemetry/smart-gateway-operator-1-build" Oct 08 00:20:19 crc kubenswrapper[4967]: I1008 00:20:19.710824 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-r7snr-pull\" (UniqueName: \"kubernetes.io/secret/f47a4120-aa3b-400b-aa4f-42b3262efd73-builder-dockercfg-r7snr-pull\") pod \"smart-gateway-operator-1-build\" (UID: \"f47a4120-aa3b-400b-aa4f-42b3262efd73\") " pod="service-telemetry/smart-gateway-operator-1-build" Oct 08 00:20:19 crc kubenswrapper[4967]: I1008 00:20:19.710911 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-r7snr-push\" (UniqueName: \"kubernetes.io/secret/f47a4120-aa3b-400b-aa4f-42b3262efd73-builder-dockercfg-r7snr-push\") pod \"smart-gateway-operator-1-build\" (UID: \"f47a4120-aa3b-400b-aa4f-42b3262efd73\") " pod="service-telemetry/smart-gateway-operator-1-build" Oct 08 00:20:19 crc kubenswrapper[4967]: I1008 00:20:19.737063 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j8rnd\" (UniqueName: \"kubernetes.io/projected/f47a4120-aa3b-400b-aa4f-42b3262efd73-kube-api-access-j8rnd\") pod \"smart-gateway-operator-1-build\" (UID: \"f47a4120-aa3b-400b-aa4f-42b3262efd73\") " pod="service-telemetry/smart-gateway-operator-1-build" Oct 08 00:20:19 crc kubenswrapper[4967]: I1008 00:20:19.828760 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-1-build" Oct 08 00:20:20 crc kubenswrapper[4967]: I1008 00:20:20.069213 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/smart-gateway-operator-1-build"] Oct 08 00:20:20 crc kubenswrapper[4967]: I1008 00:20:20.468231 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-1-build" event={"ID":"f47a4120-aa3b-400b-aa4f-42b3262efd73","Type":"ContainerStarted","Data":"78d28c3fe26a26fa10cde2fb1b2815aeaeb77ff77281fde7bafba96f13a6ca1e"} Oct 08 00:20:21 crc kubenswrapper[4967]: I1008 00:20:21.479051 4967 generic.go:334] "Generic (PLEG): container finished" podID="f47a4120-aa3b-400b-aa4f-42b3262efd73" containerID="b66ff950e0ff9eb1e5bdf877a710e93749438176d9113f9a4884802d4ee6331e" exitCode=0 Oct 08 00:20:21 crc kubenswrapper[4967]: I1008 00:20:21.479140 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-1-build" event={"ID":"f47a4120-aa3b-400b-aa4f-42b3262efd73","Type":"ContainerDied","Data":"b66ff950e0ff9eb1e5bdf877a710e93749438176d9113f9a4884802d4ee6331e"} Oct 08 00:20:22 crc kubenswrapper[4967]: I1008 00:20:22.495817 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-1-build" event={"ID":"f47a4120-aa3b-400b-aa4f-42b3262efd73","Type":"ContainerStarted","Data":"65ee9e80cdd57ed752351a8e98231fb77c1db9d8d5ab579c6aca5198e981eb48"} Oct 08 00:20:22 crc kubenswrapper[4967]: I1008 00:20:22.530631 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/smart-gateway-operator-1-build" podStartSLOduration=3.530613142 podStartE2EDuration="3.530613142s" podCreationTimestamp="2025-10-08 00:20:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:20:22.529636095 +0000 UTC m=+821.180170774" watchObservedRunningTime="2025-10-08 00:20:22.530613142 +0000 UTC m=+821.181147811" Oct 08 00:20:30 crc kubenswrapper[4967]: I1008 00:20:30.337956 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/smart-gateway-operator-1-build"] Oct 08 00:20:30 crc kubenswrapper[4967]: I1008 00:20:30.339111 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="service-telemetry/smart-gateway-operator-1-build" podUID="f47a4120-aa3b-400b-aa4f-42b3262efd73" containerName="docker-build" containerID="cri-o://65ee9e80cdd57ed752351a8e98231fb77c1db9d8d5ab579c6aca5198e981eb48" gracePeriod=30 Oct 08 00:20:31 crc kubenswrapper[4967]: I1008 00:20:31.940814 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/smart-gateway-operator-2-build"] Oct 08 00:20:31 crc kubenswrapper[4967]: I1008 00:20:31.942245 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-2-build" Oct 08 00:20:31 crc kubenswrapper[4967]: I1008 00:20:31.946651 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"smart-gateway-operator-2-ca" Oct 08 00:20:31 crc kubenswrapper[4967]: I1008 00:20:31.948356 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"smart-gateway-operator-2-sys-config" Oct 08 00:20:31 crc kubenswrapper[4967]: I1008 00:20:31.952923 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"smart-gateway-operator-2-global-ca" Oct 08 00:20:31 crc kubenswrapper[4967]: I1008 00:20:31.971388 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/smart-gateway-operator-2-build"] Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.131361 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/52664964-4de9-49b7-af7f-b97194774460-build-blob-cache\") pod \"smart-gateway-operator-2-build\" (UID: \"52664964-4de9-49b7-af7f-b97194774460\") " pod="service-telemetry/smart-gateway-operator-2-build" Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.132040 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-r7snr-pull\" (UniqueName: \"kubernetes.io/secret/52664964-4de9-49b7-af7f-b97194774460-builder-dockercfg-r7snr-pull\") pod \"smart-gateway-operator-2-build\" (UID: \"52664964-4de9-49b7-af7f-b97194774460\") " pod="service-telemetry/smart-gateway-operator-2-build" Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.132188 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kslzj\" (UniqueName: \"kubernetes.io/projected/52664964-4de9-49b7-af7f-b97194774460-kube-api-access-kslzj\") pod \"smart-gateway-operator-2-build\" (UID: \"52664964-4de9-49b7-af7f-b97194774460\") " pod="service-telemetry/smart-gateway-operator-2-build" Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.132336 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/52664964-4de9-49b7-af7f-b97194774460-container-storage-run\") pod \"smart-gateway-operator-2-build\" (UID: \"52664964-4de9-49b7-af7f-b97194774460\") " pod="service-telemetry/smart-gateway-operator-2-build" Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.132447 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-r7snr-push\" (UniqueName: \"kubernetes.io/secret/52664964-4de9-49b7-af7f-b97194774460-builder-dockercfg-r7snr-push\") pod \"smart-gateway-operator-2-build\" (UID: \"52664964-4de9-49b7-af7f-b97194774460\") " pod="service-telemetry/smart-gateway-operator-2-build" Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.132559 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/52664964-4de9-49b7-af7f-b97194774460-node-pullsecrets\") pod \"smart-gateway-operator-2-build\" (UID: \"52664964-4de9-49b7-af7f-b97194774460\") " pod="service-telemetry/smart-gateway-operator-2-build" Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.132704 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/52664964-4de9-49b7-af7f-b97194774460-build-ca-bundles\") pod \"smart-gateway-operator-2-build\" (UID: \"52664964-4de9-49b7-af7f-b97194774460\") " pod="service-telemetry/smart-gateway-operator-2-build" Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.132831 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/52664964-4de9-49b7-af7f-b97194774460-buildworkdir\") pod \"smart-gateway-operator-2-build\" (UID: \"52664964-4de9-49b7-af7f-b97194774460\") " pod="service-telemetry/smart-gateway-operator-2-build" Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.132952 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/52664964-4de9-49b7-af7f-b97194774460-build-proxy-ca-bundles\") pod \"smart-gateway-operator-2-build\" (UID: \"52664964-4de9-49b7-af7f-b97194774460\") " pod="service-telemetry/smart-gateway-operator-2-build" Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.133054 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/52664964-4de9-49b7-af7f-b97194774460-build-system-configs\") pod \"smart-gateway-operator-2-build\" (UID: \"52664964-4de9-49b7-af7f-b97194774460\") " pod="service-telemetry/smart-gateway-operator-2-build" Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.133158 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/52664964-4de9-49b7-af7f-b97194774460-container-storage-root\") pod \"smart-gateway-operator-2-build\" (UID: \"52664964-4de9-49b7-af7f-b97194774460\") " pod="service-telemetry/smart-gateway-operator-2-build" Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.133278 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/52664964-4de9-49b7-af7f-b97194774460-buildcachedir\") pod \"smart-gateway-operator-2-build\" (UID: \"52664964-4de9-49b7-af7f-b97194774460\") " pod="service-telemetry/smart-gateway-operator-2-build" Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.235298 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/52664964-4de9-49b7-af7f-b97194774460-build-proxy-ca-bundles\") pod \"smart-gateway-operator-2-build\" (UID: \"52664964-4de9-49b7-af7f-b97194774460\") " pod="service-telemetry/smart-gateway-operator-2-build" Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.235827 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/52664964-4de9-49b7-af7f-b97194774460-build-system-configs\") pod \"smart-gateway-operator-2-build\" (UID: \"52664964-4de9-49b7-af7f-b97194774460\") " pod="service-telemetry/smart-gateway-operator-2-build" Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.236098 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/52664964-4de9-49b7-af7f-b97194774460-container-storage-root\") pod \"smart-gateway-operator-2-build\" (UID: \"52664964-4de9-49b7-af7f-b97194774460\") " pod="service-telemetry/smart-gateway-operator-2-build" Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.236345 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/52664964-4de9-49b7-af7f-b97194774460-buildcachedir\") pod \"smart-gateway-operator-2-build\" (UID: \"52664964-4de9-49b7-af7f-b97194774460\") " pod="service-telemetry/smart-gateway-operator-2-build" Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.236599 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/52664964-4de9-49b7-af7f-b97194774460-build-system-configs\") pod \"smart-gateway-operator-2-build\" (UID: \"52664964-4de9-49b7-af7f-b97194774460\") " pod="service-telemetry/smart-gateway-operator-2-build" Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.236475 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/52664964-4de9-49b7-af7f-b97194774460-buildcachedir\") pod \"smart-gateway-operator-2-build\" (UID: \"52664964-4de9-49b7-af7f-b97194774460\") " pod="service-telemetry/smart-gateway-operator-2-build" Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.236890 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/52664964-4de9-49b7-af7f-b97194774460-build-proxy-ca-bundles\") pod \"smart-gateway-operator-2-build\" (UID: \"52664964-4de9-49b7-af7f-b97194774460\") " pod="service-telemetry/smart-gateway-operator-2-build" Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.236896 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/52664964-4de9-49b7-af7f-b97194774460-container-storage-root\") pod \"smart-gateway-operator-2-build\" (UID: \"52664964-4de9-49b7-af7f-b97194774460\") " pod="service-telemetry/smart-gateway-operator-2-build" Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.237226 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/52664964-4de9-49b7-af7f-b97194774460-build-blob-cache\") pod \"smart-gateway-operator-2-build\" (UID: \"52664964-4de9-49b7-af7f-b97194774460\") " pod="service-telemetry/smart-gateway-operator-2-build" Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.237546 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-r7snr-pull\" (UniqueName: \"kubernetes.io/secret/52664964-4de9-49b7-af7f-b97194774460-builder-dockercfg-r7snr-pull\") pod \"smart-gateway-operator-2-build\" (UID: \"52664964-4de9-49b7-af7f-b97194774460\") " pod="service-telemetry/smart-gateway-operator-2-build" Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.237973 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kslzj\" (UniqueName: \"kubernetes.io/projected/52664964-4de9-49b7-af7f-b97194774460-kube-api-access-kslzj\") pod \"smart-gateway-operator-2-build\" (UID: \"52664964-4de9-49b7-af7f-b97194774460\") " pod="service-telemetry/smart-gateway-operator-2-build" Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.238224 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-r7snr-push\" (UniqueName: \"kubernetes.io/secret/52664964-4de9-49b7-af7f-b97194774460-builder-dockercfg-r7snr-push\") pod \"smart-gateway-operator-2-build\" (UID: \"52664964-4de9-49b7-af7f-b97194774460\") " pod="service-telemetry/smart-gateway-operator-2-build" Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.238451 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/52664964-4de9-49b7-af7f-b97194774460-container-storage-run\") pod \"smart-gateway-operator-2-build\" (UID: \"52664964-4de9-49b7-af7f-b97194774460\") " pod="service-telemetry/smart-gateway-operator-2-build" Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.238877 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/52664964-4de9-49b7-af7f-b97194774460-node-pullsecrets\") pod \"smart-gateway-operator-2-build\" (UID: \"52664964-4de9-49b7-af7f-b97194774460\") " pod="service-telemetry/smart-gateway-operator-2-build" Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.237752 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/52664964-4de9-49b7-af7f-b97194774460-build-blob-cache\") pod \"smart-gateway-operator-2-build\" (UID: \"52664964-4de9-49b7-af7f-b97194774460\") " pod="service-telemetry/smart-gateway-operator-2-build" Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.238965 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/52664964-4de9-49b7-af7f-b97194774460-node-pullsecrets\") pod \"smart-gateway-operator-2-build\" (UID: \"52664964-4de9-49b7-af7f-b97194774460\") " pod="service-telemetry/smart-gateway-operator-2-build" Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.239183 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/52664964-4de9-49b7-af7f-b97194774460-container-storage-run\") pod \"smart-gateway-operator-2-build\" (UID: \"52664964-4de9-49b7-af7f-b97194774460\") " pod="service-telemetry/smart-gateway-operator-2-build" Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.239052 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/52664964-4de9-49b7-af7f-b97194774460-build-ca-bundles\") pod \"smart-gateway-operator-2-build\" (UID: \"52664964-4de9-49b7-af7f-b97194774460\") " pod="service-telemetry/smart-gateway-operator-2-build" Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.239343 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/52664964-4de9-49b7-af7f-b97194774460-buildworkdir\") pod \"smart-gateway-operator-2-build\" (UID: \"52664964-4de9-49b7-af7f-b97194774460\") " pod="service-telemetry/smart-gateway-operator-2-build" Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.240132 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/52664964-4de9-49b7-af7f-b97194774460-buildworkdir\") pod \"smart-gateway-operator-2-build\" (UID: \"52664964-4de9-49b7-af7f-b97194774460\") " pod="service-telemetry/smart-gateway-operator-2-build" Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.241283 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/52664964-4de9-49b7-af7f-b97194774460-build-ca-bundles\") pod \"smart-gateway-operator-2-build\" (UID: \"52664964-4de9-49b7-af7f-b97194774460\") " pod="service-telemetry/smart-gateway-operator-2-build" Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.247591 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-r7snr-pull\" (UniqueName: \"kubernetes.io/secret/52664964-4de9-49b7-af7f-b97194774460-builder-dockercfg-r7snr-pull\") pod \"smart-gateway-operator-2-build\" (UID: \"52664964-4de9-49b7-af7f-b97194774460\") " pod="service-telemetry/smart-gateway-operator-2-build" Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.247600 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-r7snr-push\" (UniqueName: \"kubernetes.io/secret/52664964-4de9-49b7-af7f-b97194774460-builder-dockercfg-r7snr-push\") pod \"smart-gateway-operator-2-build\" (UID: \"52664964-4de9-49b7-af7f-b97194774460\") " pod="service-telemetry/smart-gateway-operator-2-build" Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.263578 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kslzj\" (UniqueName: \"kubernetes.io/projected/52664964-4de9-49b7-af7f-b97194774460-kube-api-access-kslzj\") pod \"smart-gateway-operator-2-build\" (UID: \"52664964-4de9-49b7-af7f-b97194774460\") " pod="service-telemetry/smart-gateway-operator-2-build" Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.554571 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_smart-gateway-operator-1-build_f47a4120-aa3b-400b-aa4f-42b3262efd73/docker-build/0.log" Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.555499 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-1-build" Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.561111 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-2-build" Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.572405 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_smart-gateway-operator-1-build_f47a4120-aa3b-400b-aa4f-42b3262efd73/docker-build/0.log" Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.572899 4967 generic.go:334] "Generic (PLEG): container finished" podID="f47a4120-aa3b-400b-aa4f-42b3262efd73" containerID="65ee9e80cdd57ed752351a8e98231fb77c1db9d8d5ab579c6aca5198e981eb48" exitCode=1 Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.572943 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-1-build" event={"ID":"f47a4120-aa3b-400b-aa4f-42b3262efd73","Type":"ContainerDied","Data":"65ee9e80cdd57ed752351a8e98231fb77c1db9d8d5ab579c6aca5198e981eb48"} Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.572971 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-1-build" event={"ID":"f47a4120-aa3b-400b-aa4f-42b3262efd73","Type":"ContainerDied","Data":"78d28c3fe26a26fa10cde2fb1b2815aeaeb77ff77281fde7bafba96f13a6ca1e"} Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.572992 4967 scope.go:117] "RemoveContainer" containerID="65ee9e80cdd57ed752351a8e98231fb77c1db9d8d5ab579c6aca5198e981eb48" Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.573262 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-1-build" Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.643543 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j8rnd\" (UniqueName: \"kubernetes.io/projected/f47a4120-aa3b-400b-aa4f-42b3262efd73-kube-api-access-j8rnd\") pod \"f47a4120-aa3b-400b-aa4f-42b3262efd73\" (UID: \"f47a4120-aa3b-400b-aa4f-42b3262efd73\") " Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.643588 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-r7snr-pull\" (UniqueName: \"kubernetes.io/secret/f47a4120-aa3b-400b-aa4f-42b3262efd73-builder-dockercfg-r7snr-pull\") pod \"f47a4120-aa3b-400b-aa4f-42b3262efd73\" (UID: \"f47a4120-aa3b-400b-aa4f-42b3262efd73\") " Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.643612 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/f47a4120-aa3b-400b-aa4f-42b3262efd73-build-system-configs\") pod \"f47a4120-aa3b-400b-aa4f-42b3262efd73\" (UID: \"f47a4120-aa3b-400b-aa4f-42b3262efd73\") " Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.644384 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f47a4120-aa3b-400b-aa4f-42b3262efd73-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "f47a4120-aa3b-400b-aa4f-42b3262efd73" (UID: "f47a4120-aa3b-400b-aa4f-42b3262efd73"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.650117 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f47a4120-aa3b-400b-aa4f-42b3262efd73-builder-dockercfg-r7snr-pull" (OuterVolumeSpecName: "builder-dockercfg-r7snr-pull") pod "f47a4120-aa3b-400b-aa4f-42b3262efd73" (UID: "f47a4120-aa3b-400b-aa4f-42b3262efd73"). InnerVolumeSpecName "builder-dockercfg-r7snr-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.650708 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f47a4120-aa3b-400b-aa4f-42b3262efd73-kube-api-access-j8rnd" (OuterVolumeSpecName: "kube-api-access-j8rnd") pod "f47a4120-aa3b-400b-aa4f-42b3262efd73" (UID: "f47a4120-aa3b-400b-aa4f-42b3262efd73"). InnerVolumeSpecName "kube-api-access-j8rnd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.669566 4967 scope.go:117] "RemoveContainer" containerID="b66ff950e0ff9eb1e5bdf877a710e93749438176d9113f9a4884802d4ee6331e" Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.693370 4967 scope.go:117] "RemoveContainer" containerID="65ee9e80cdd57ed752351a8e98231fb77c1db9d8d5ab579c6aca5198e981eb48" Oct 08 00:20:32 crc kubenswrapper[4967]: E1008 00:20:32.693913 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"65ee9e80cdd57ed752351a8e98231fb77c1db9d8d5ab579c6aca5198e981eb48\": container with ID starting with 65ee9e80cdd57ed752351a8e98231fb77c1db9d8d5ab579c6aca5198e981eb48 not found: ID does not exist" containerID="65ee9e80cdd57ed752351a8e98231fb77c1db9d8d5ab579c6aca5198e981eb48" Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.693979 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65ee9e80cdd57ed752351a8e98231fb77c1db9d8d5ab579c6aca5198e981eb48"} err="failed to get container status \"65ee9e80cdd57ed752351a8e98231fb77c1db9d8d5ab579c6aca5198e981eb48\": rpc error: code = NotFound desc = could not find container \"65ee9e80cdd57ed752351a8e98231fb77c1db9d8d5ab579c6aca5198e981eb48\": container with ID starting with 65ee9e80cdd57ed752351a8e98231fb77c1db9d8d5ab579c6aca5198e981eb48 not found: ID does not exist" Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.694046 4967 scope.go:117] "RemoveContainer" containerID="b66ff950e0ff9eb1e5bdf877a710e93749438176d9113f9a4884802d4ee6331e" Oct 08 00:20:32 crc kubenswrapper[4967]: E1008 00:20:32.694470 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b66ff950e0ff9eb1e5bdf877a710e93749438176d9113f9a4884802d4ee6331e\": container with ID starting with b66ff950e0ff9eb1e5bdf877a710e93749438176d9113f9a4884802d4ee6331e not found: ID does not exist" containerID="b66ff950e0ff9eb1e5bdf877a710e93749438176d9113f9a4884802d4ee6331e" Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.694497 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b66ff950e0ff9eb1e5bdf877a710e93749438176d9113f9a4884802d4ee6331e"} err="failed to get container status \"b66ff950e0ff9eb1e5bdf877a710e93749438176d9113f9a4884802d4ee6331e\": rpc error: code = NotFound desc = could not find container \"b66ff950e0ff9eb1e5bdf877a710e93749438176d9113f9a4884802d4ee6331e\": container with ID starting with b66ff950e0ff9eb1e5bdf877a710e93749438176d9113f9a4884802d4ee6331e not found: ID does not exist" Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.744558 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/f47a4120-aa3b-400b-aa4f-42b3262efd73-build-blob-cache\") pod \"f47a4120-aa3b-400b-aa4f-42b3262efd73\" (UID: \"f47a4120-aa3b-400b-aa4f-42b3262efd73\") " Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.744702 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-r7snr-push\" (UniqueName: \"kubernetes.io/secret/f47a4120-aa3b-400b-aa4f-42b3262efd73-builder-dockercfg-r7snr-push\") pod \"f47a4120-aa3b-400b-aa4f-42b3262efd73\" (UID: \"f47a4120-aa3b-400b-aa4f-42b3262efd73\") " Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.744759 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f47a4120-aa3b-400b-aa4f-42b3262efd73-build-ca-bundles\") pod \"f47a4120-aa3b-400b-aa4f-42b3262efd73\" (UID: \"f47a4120-aa3b-400b-aa4f-42b3262efd73\") " Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.744817 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/f47a4120-aa3b-400b-aa4f-42b3262efd73-node-pullsecrets\") pod \"f47a4120-aa3b-400b-aa4f-42b3262efd73\" (UID: \"f47a4120-aa3b-400b-aa4f-42b3262efd73\") " Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.744854 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/f47a4120-aa3b-400b-aa4f-42b3262efd73-container-storage-root\") pod \"f47a4120-aa3b-400b-aa4f-42b3262efd73\" (UID: \"f47a4120-aa3b-400b-aa4f-42b3262efd73\") " Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.744902 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f47a4120-aa3b-400b-aa4f-42b3262efd73-build-proxy-ca-bundles\") pod \"f47a4120-aa3b-400b-aa4f-42b3262efd73\" (UID: \"f47a4120-aa3b-400b-aa4f-42b3262efd73\") " Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.744937 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/f47a4120-aa3b-400b-aa4f-42b3262efd73-buildcachedir\") pod \"f47a4120-aa3b-400b-aa4f-42b3262efd73\" (UID: \"f47a4120-aa3b-400b-aa4f-42b3262efd73\") " Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.744997 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/f47a4120-aa3b-400b-aa4f-42b3262efd73-buildworkdir\") pod \"f47a4120-aa3b-400b-aa4f-42b3262efd73\" (UID: \"f47a4120-aa3b-400b-aa4f-42b3262efd73\") " Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.745042 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/f47a4120-aa3b-400b-aa4f-42b3262efd73-container-storage-run\") pod \"f47a4120-aa3b-400b-aa4f-42b3262efd73\" (UID: \"f47a4120-aa3b-400b-aa4f-42b3262efd73\") " Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.745460 4967 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/f47a4120-aa3b-400b-aa4f-42b3262efd73-build-system-configs\") on node \"crc\" DevicePath \"\"" Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.745493 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j8rnd\" (UniqueName: \"kubernetes.io/projected/f47a4120-aa3b-400b-aa4f-42b3262efd73-kube-api-access-j8rnd\") on node \"crc\" DevicePath \"\"" Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.745511 4967 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-r7snr-pull\" (UniqueName: \"kubernetes.io/secret/f47a4120-aa3b-400b-aa4f-42b3262efd73-builder-dockercfg-r7snr-pull\") on node \"crc\" DevicePath \"\"" Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.745873 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f47a4120-aa3b-400b-aa4f-42b3262efd73-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "f47a4120-aa3b-400b-aa4f-42b3262efd73" (UID: "f47a4120-aa3b-400b-aa4f-42b3262efd73"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.745873 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f47a4120-aa3b-400b-aa4f-42b3262efd73-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "f47a4120-aa3b-400b-aa4f-42b3262efd73" (UID: "f47a4120-aa3b-400b-aa4f-42b3262efd73"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.745935 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f47a4120-aa3b-400b-aa4f-42b3262efd73-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "f47a4120-aa3b-400b-aa4f-42b3262efd73" (UID: "f47a4120-aa3b-400b-aa4f-42b3262efd73"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.745986 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f47a4120-aa3b-400b-aa4f-42b3262efd73-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "f47a4120-aa3b-400b-aa4f-42b3262efd73" (UID: "f47a4120-aa3b-400b-aa4f-42b3262efd73"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.746460 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f47a4120-aa3b-400b-aa4f-42b3262efd73-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "f47a4120-aa3b-400b-aa4f-42b3262efd73" (UID: "f47a4120-aa3b-400b-aa4f-42b3262efd73"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.747047 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f47a4120-aa3b-400b-aa4f-42b3262efd73-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "f47a4120-aa3b-400b-aa4f-42b3262efd73" (UID: "f47a4120-aa3b-400b-aa4f-42b3262efd73"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.748354 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f47a4120-aa3b-400b-aa4f-42b3262efd73-builder-dockercfg-r7snr-push" (OuterVolumeSpecName: "builder-dockercfg-r7snr-push") pod "f47a4120-aa3b-400b-aa4f-42b3262efd73" (UID: "f47a4120-aa3b-400b-aa4f-42b3262efd73"). InnerVolumeSpecName "builder-dockercfg-r7snr-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.805247 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/smart-gateway-operator-2-build"] Oct 08 00:20:32 crc kubenswrapper[4967]: W1008 00:20:32.811960 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod52664964_4de9_49b7_af7f_b97194774460.slice/crio-04f5ec5005e7aa5c6dc2246bd3e8a13bcaddb21a468a8ff83c84cfa122411dc1 WatchSource:0}: Error finding container 04f5ec5005e7aa5c6dc2246bd3e8a13bcaddb21a468a8ff83c84cfa122411dc1: Status 404 returned error can't find the container with id 04f5ec5005e7aa5c6dc2246bd3e8a13bcaddb21a468a8ff83c84cfa122411dc1 Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.846716 4967 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/f47a4120-aa3b-400b-aa4f-42b3262efd73-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.846748 4967 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f47a4120-aa3b-400b-aa4f-42b3262efd73-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.846760 4967 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/f47a4120-aa3b-400b-aa4f-42b3262efd73-buildcachedir\") on node \"crc\" DevicePath \"\"" Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.846776 4967 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/f47a4120-aa3b-400b-aa4f-42b3262efd73-buildworkdir\") on node \"crc\" DevicePath \"\"" Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.846786 4967 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/f47a4120-aa3b-400b-aa4f-42b3262efd73-container-storage-run\") on node \"crc\" DevicePath \"\"" Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.846794 4967 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-r7snr-push\" (UniqueName: \"kubernetes.io/secret/f47a4120-aa3b-400b-aa4f-42b3262efd73-builder-dockercfg-r7snr-push\") on node \"crc\" DevicePath \"\"" Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.846802 4967 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f47a4120-aa3b-400b-aa4f-42b3262efd73-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.930835 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f47a4120-aa3b-400b-aa4f-42b3262efd73-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "f47a4120-aa3b-400b-aa4f-42b3262efd73" (UID: "f47a4120-aa3b-400b-aa4f-42b3262efd73"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:20:32 crc kubenswrapper[4967]: I1008 00:20:32.948901 4967 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/f47a4120-aa3b-400b-aa4f-42b3262efd73-build-blob-cache\") on node \"crc\" DevicePath \"\"" Oct 08 00:20:33 crc kubenswrapper[4967]: I1008 00:20:33.252948 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f47a4120-aa3b-400b-aa4f-42b3262efd73-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "f47a4120-aa3b-400b-aa4f-42b3262efd73" (UID: "f47a4120-aa3b-400b-aa4f-42b3262efd73"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:20:33 crc kubenswrapper[4967]: I1008 00:20:33.355601 4967 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/f47a4120-aa3b-400b-aa4f-42b3262efd73-container-storage-root\") on node \"crc\" DevicePath \"\"" Oct 08 00:20:33 crc kubenswrapper[4967]: I1008 00:20:33.505081 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/smart-gateway-operator-1-build"] Oct 08 00:20:33 crc kubenswrapper[4967]: I1008 00:20:33.510385 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["service-telemetry/smart-gateway-operator-1-build"] Oct 08 00:20:33 crc kubenswrapper[4967]: I1008 00:20:33.581364 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-2-build" event={"ID":"52664964-4de9-49b7-af7f-b97194774460","Type":"ContainerStarted","Data":"2ed0d1c17ded4726d9a7d2353592919eb1f0f88b42e3573b7a358d38ddd69690"} Oct 08 00:20:33 crc kubenswrapper[4967]: I1008 00:20:33.581426 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-2-build" event={"ID":"52664964-4de9-49b7-af7f-b97194774460","Type":"ContainerStarted","Data":"04f5ec5005e7aa5c6dc2246bd3e8a13bcaddb21a468a8ff83c84cfa122411dc1"} Oct 08 00:20:34 crc kubenswrapper[4967]: I1008 00:20:34.130530 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f47a4120-aa3b-400b-aa4f-42b3262efd73" path="/var/lib/kubelet/pods/f47a4120-aa3b-400b-aa4f-42b3262efd73/volumes" Oct 08 00:20:34 crc kubenswrapper[4967]: I1008 00:20:34.591842 4967 generic.go:334] "Generic (PLEG): container finished" podID="52664964-4de9-49b7-af7f-b97194774460" containerID="2ed0d1c17ded4726d9a7d2353592919eb1f0f88b42e3573b7a358d38ddd69690" exitCode=0 Oct 08 00:20:34 crc kubenswrapper[4967]: I1008 00:20:34.591897 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-2-build" event={"ID":"52664964-4de9-49b7-af7f-b97194774460","Type":"ContainerDied","Data":"2ed0d1c17ded4726d9a7d2353592919eb1f0f88b42e3573b7a358d38ddd69690"} Oct 08 00:20:35 crc kubenswrapper[4967]: I1008 00:20:35.603219 4967 generic.go:334] "Generic (PLEG): container finished" podID="52664964-4de9-49b7-af7f-b97194774460" containerID="8f36b745bc4b608857b8251551db69a6512250e4b0fd20afa5832107e46f00b8" exitCode=0 Oct 08 00:20:35 crc kubenswrapper[4967]: I1008 00:20:35.603325 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-2-build" event={"ID":"52664964-4de9-49b7-af7f-b97194774460","Type":"ContainerDied","Data":"8f36b745bc4b608857b8251551db69a6512250e4b0fd20afa5832107e46f00b8"} Oct 08 00:20:35 crc kubenswrapper[4967]: I1008 00:20:35.646902 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_smart-gateway-operator-2-build_52664964-4de9-49b7-af7f-b97194774460/manage-dockerfile/0.log" Oct 08 00:20:36 crc kubenswrapper[4967]: I1008 00:20:36.619720 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-2-build" event={"ID":"52664964-4de9-49b7-af7f-b97194774460","Type":"ContainerStarted","Data":"eda9082c24912a8da64ef34e651bbd97bd2d63594c29f03b06a61a8b4cf14650"} Oct 08 00:20:36 crc kubenswrapper[4967]: I1008 00:20:36.660257 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/smart-gateway-operator-2-build" podStartSLOduration=5.660234243 podStartE2EDuration="5.660234243s" podCreationTimestamp="2025-10-08 00:20:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:20:36.652988937 +0000 UTC m=+835.303523586" watchObservedRunningTime="2025-10-08 00:20:36.660234243 +0000 UTC m=+835.310768922" Oct 08 00:20:36 crc kubenswrapper[4967]: I1008 00:20:36.919973 4967 patch_prober.go:28] interesting pod/machine-config-daemon-xvdtp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 00:20:36 crc kubenswrapper[4967]: I1008 00:20:36.920034 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" podUID="9461162a-c27d-465e-9ca7-2d3936877cd8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 00:20:36 crc kubenswrapper[4967]: I1008 00:20:36.920078 4967 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" Oct 08 00:20:36 crc kubenswrapper[4967]: I1008 00:20:36.920570 4967 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"312ed5e75849b65841b141a338e52c38e6c54f84a2e110368c62a60516be0c6b"} pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 00:20:36 crc kubenswrapper[4967]: I1008 00:20:36.920626 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" podUID="9461162a-c27d-465e-9ca7-2d3936877cd8" containerName="machine-config-daemon" containerID="cri-o://312ed5e75849b65841b141a338e52c38e6c54f84a2e110368c62a60516be0c6b" gracePeriod=600 Oct 08 00:20:37 crc kubenswrapper[4967]: I1008 00:20:37.632203 4967 generic.go:334] "Generic (PLEG): container finished" podID="9461162a-c27d-465e-9ca7-2d3936877cd8" containerID="312ed5e75849b65841b141a338e52c38e6c54f84a2e110368c62a60516be0c6b" exitCode=0 Oct 08 00:20:37 crc kubenswrapper[4967]: I1008 00:20:37.632281 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" event={"ID":"9461162a-c27d-465e-9ca7-2d3936877cd8","Type":"ContainerDied","Data":"312ed5e75849b65841b141a338e52c38e6c54f84a2e110368c62a60516be0c6b"} Oct 08 00:20:37 crc kubenswrapper[4967]: I1008 00:20:37.632535 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" event={"ID":"9461162a-c27d-465e-9ca7-2d3936877cd8","Type":"ContainerStarted","Data":"bd8678b8254640836f0db760c5c6a0da933718c28fc5033465f9875f7bd725b1"} Oct 08 00:20:37 crc kubenswrapper[4967]: I1008 00:20:37.632555 4967 scope.go:117] "RemoveContainer" containerID="f9c94f180083f2a4c120cfed331389932c804613b3fbe7d131d60b469072cb08" Oct 08 00:21:51 crc kubenswrapper[4967]: I1008 00:21:51.192364 4967 generic.go:334] "Generic (PLEG): container finished" podID="52664964-4de9-49b7-af7f-b97194774460" containerID="eda9082c24912a8da64ef34e651bbd97bd2d63594c29f03b06a61a8b4cf14650" exitCode=0 Oct 08 00:21:51 crc kubenswrapper[4967]: I1008 00:21:51.192470 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-2-build" event={"ID":"52664964-4de9-49b7-af7f-b97194774460","Type":"ContainerDied","Data":"eda9082c24912a8da64ef34e651bbd97bd2d63594c29f03b06a61a8b4cf14650"} Oct 08 00:21:52 crc kubenswrapper[4967]: I1008 00:21:52.555829 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-2-build" Oct 08 00:21:52 crc kubenswrapper[4967]: I1008 00:21:52.666984 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/52664964-4de9-49b7-af7f-b97194774460-build-blob-cache\") pod \"52664964-4de9-49b7-af7f-b97194774460\" (UID: \"52664964-4de9-49b7-af7f-b97194774460\") " Oct 08 00:21:52 crc kubenswrapper[4967]: I1008 00:21:52.667050 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/52664964-4de9-49b7-af7f-b97194774460-build-proxy-ca-bundles\") pod \"52664964-4de9-49b7-af7f-b97194774460\" (UID: \"52664964-4de9-49b7-af7f-b97194774460\") " Oct 08 00:21:52 crc kubenswrapper[4967]: I1008 00:21:52.667098 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/52664964-4de9-49b7-af7f-b97194774460-buildcachedir\") pod \"52664964-4de9-49b7-af7f-b97194774460\" (UID: \"52664964-4de9-49b7-af7f-b97194774460\") " Oct 08 00:21:52 crc kubenswrapper[4967]: I1008 00:21:52.667127 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/52664964-4de9-49b7-af7f-b97194774460-build-system-configs\") pod \"52664964-4de9-49b7-af7f-b97194774460\" (UID: \"52664964-4de9-49b7-af7f-b97194774460\") " Oct 08 00:21:52 crc kubenswrapper[4967]: I1008 00:21:52.667160 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/52664964-4de9-49b7-af7f-b97194774460-container-storage-root\") pod \"52664964-4de9-49b7-af7f-b97194774460\" (UID: \"52664964-4de9-49b7-af7f-b97194774460\") " Oct 08 00:21:52 crc kubenswrapper[4967]: I1008 00:21:52.667188 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/52664964-4de9-49b7-af7f-b97194774460-container-storage-run\") pod \"52664964-4de9-49b7-af7f-b97194774460\" (UID: \"52664964-4de9-49b7-af7f-b97194774460\") " Oct 08 00:21:52 crc kubenswrapper[4967]: I1008 00:21:52.667211 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kslzj\" (UniqueName: \"kubernetes.io/projected/52664964-4de9-49b7-af7f-b97194774460-kube-api-access-kslzj\") pod \"52664964-4de9-49b7-af7f-b97194774460\" (UID: \"52664964-4de9-49b7-af7f-b97194774460\") " Oct 08 00:21:52 crc kubenswrapper[4967]: I1008 00:21:52.667215 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/52664964-4de9-49b7-af7f-b97194774460-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "52664964-4de9-49b7-af7f-b97194774460" (UID: "52664964-4de9-49b7-af7f-b97194774460"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 00:21:52 crc kubenswrapper[4967]: I1008 00:21:52.667251 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-r7snr-pull\" (UniqueName: \"kubernetes.io/secret/52664964-4de9-49b7-af7f-b97194774460-builder-dockercfg-r7snr-pull\") pod \"52664964-4de9-49b7-af7f-b97194774460\" (UID: \"52664964-4de9-49b7-af7f-b97194774460\") " Oct 08 00:21:52 crc kubenswrapper[4967]: I1008 00:21:52.667274 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/52664964-4de9-49b7-af7f-b97194774460-buildworkdir\") pod \"52664964-4de9-49b7-af7f-b97194774460\" (UID: \"52664964-4de9-49b7-af7f-b97194774460\") " Oct 08 00:21:52 crc kubenswrapper[4967]: I1008 00:21:52.667308 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/52664964-4de9-49b7-af7f-b97194774460-node-pullsecrets\") pod \"52664964-4de9-49b7-af7f-b97194774460\" (UID: \"52664964-4de9-49b7-af7f-b97194774460\") " Oct 08 00:21:52 crc kubenswrapper[4967]: I1008 00:21:52.667373 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/52664964-4de9-49b7-af7f-b97194774460-build-ca-bundles\") pod \"52664964-4de9-49b7-af7f-b97194774460\" (UID: \"52664964-4de9-49b7-af7f-b97194774460\") " Oct 08 00:21:52 crc kubenswrapper[4967]: I1008 00:21:52.667395 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-r7snr-push\" (UniqueName: \"kubernetes.io/secret/52664964-4de9-49b7-af7f-b97194774460-builder-dockercfg-r7snr-push\") pod \"52664964-4de9-49b7-af7f-b97194774460\" (UID: \"52664964-4de9-49b7-af7f-b97194774460\") " Oct 08 00:21:52 crc kubenswrapper[4967]: I1008 00:21:52.667811 4967 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/52664964-4de9-49b7-af7f-b97194774460-buildcachedir\") on node \"crc\" DevicePath \"\"" Oct 08 00:21:52 crc kubenswrapper[4967]: I1008 00:21:52.668226 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/52664964-4de9-49b7-af7f-b97194774460-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "52664964-4de9-49b7-af7f-b97194774460" (UID: "52664964-4de9-49b7-af7f-b97194774460"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:21:52 crc kubenswrapper[4967]: I1008 00:21:52.668299 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/52664964-4de9-49b7-af7f-b97194774460-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "52664964-4de9-49b7-af7f-b97194774460" (UID: "52664964-4de9-49b7-af7f-b97194774460"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 00:21:52 crc kubenswrapper[4967]: I1008 00:21:52.669750 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/52664964-4de9-49b7-af7f-b97194774460-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "52664964-4de9-49b7-af7f-b97194774460" (UID: "52664964-4de9-49b7-af7f-b97194774460"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:21:52 crc kubenswrapper[4967]: I1008 00:21:52.669849 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/52664964-4de9-49b7-af7f-b97194774460-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "52664964-4de9-49b7-af7f-b97194774460" (UID: "52664964-4de9-49b7-af7f-b97194774460"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:21:52 crc kubenswrapper[4967]: I1008 00:21:52.670393 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/52664964-4de9-49b7-af7f-b97194774460-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "52664964-4de9-49b7-af7f-b97194774460" (UID: "52664964-4de9-49b7-af7f-b97194774460"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:21:52 crc kubenswrapper[4967]: I1008 00:21:52.673178 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52664964-4de9-49b7-af7f-b97194774460-builder-dockercfg-r7snr-pull" (OuterVolumeSpecName: "builder-dockercfg-r7snr-pull") pod "52664964-4de9-49b7-af7f-b97194774460" (UID: "52664964-4de9-49b7-af7f-b97194774460"). InnerVolumeSpecName "builder-dockercfg-r7snr-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:21:52 crc kubenswrapper[4967]: I1008 00:21:52.673316 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/52664964-4de9-49b7-af7f-b97194774460-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "52664964-4de9-49b7-af7f-b97194774460" (UID: "52664964-4de9-49b7-af7f-b97194774460"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:21:52 crc kubenswrapper[4967]: I1008 00:21:52.673519 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52664964-4de9-49b7-af7f-b97194774460-builder-dockercfg-r7snr-push" (OuterVolumeSpecName: "builder-dockercfg-r7snr-push") pod "52664964-4de9-49b7-af7f-b97194774460" (UID: "52664964-4de9-49b7-af7f-b97194774460"). InnerVolumeSpecName "builder-dockercfg-r7snr-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:21:52 crc kubenswrapper[4967]: I1008 00:21:52.681196 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52664964-4de9-49b7-af7f-b97194774460-kube-api-access-kslzj" (OuterVolumeSpecName: "kube-api-access-kslzj") pod "52664964-4de9-49b7-af7f-b97194774460" (UID: "52664964-4de9-49b7-af7f-b97194774460"). InnerVolumeSpecName "kube-api-access-kslzj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:21:52 crc kubenswrapper[4967]: I1008 00:21:52.769487 4967 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/52664964-4de9-49b7-af7f-b97194774460-build-system-configs\") on node \"crc\" DevicePath \"\"" Oct 08 00:21:52 crc kubenswrapper[4967]: I1008 00:21:52.769525 4967 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/52664964-4de9-49b7-af7f-b97194774460-container-storage-run\") on node \"crc\" DevicePath \"\"" Oct 08 00:21:52 crc kubenswrapper[4967]: I1008 00:21:52.769534 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kslzj\" (UniqueName: \"kubernetes.io/projected/52664964-4de9-49b7-af7f-b97194774460-kube-api-access-kslzj\") on node \"crc\" DevicePath \"\"" Oct 08 00:21:52 crc kubenswrapper[4967]: I1008 00:21:52.769542 4967 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-r7snr-pull\" (UniqueName: \"kubernetes.io/secret/52664964-4de9-49b7-af7f-b97194774460-builder-dockercfg-r7snr-pull\") on node \"crc\" DevicePath \"\"" Oct 08 00:21:52 crc kubenswrapper[4967]: I1008 00:21:52.769551 4967 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/52664964-4de9-49b7-af7f-b97194774460-buildworkdir\") on node \"crc\" DevicePath \"\"" Oct 08 00:21:52 crc kubenswrapper[4967]: I1008 00:21:52.769586 4967 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/52664964-4de9-49b7-af7f-b97194774460-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Oct 08 00:21:52 crc kubenswrapper[4967]: I1008 00:21:52.769597 4967 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/52664964-4de9-49b7-af7f-b97194774460-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 08 00:21:52 crc kubenswrapper[4967]: I1008 00:21:52.769605 4967 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-r7snr-push\" (UniqueName: \"kubernetes.io/secret/52664964-4de9-49b7-af7f-b97194774460-builder-dockercfg-r7snr-push\") on node \"crc\" DevicePath \"\"" Oct 08 00:21:52 crc kubenswrapper[4967]: I1008 00:21:52.769613 4967 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/52664964-4de9-49b7-af7f-b97194774460-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 08 00:21:52 crc kubenswrapper[4967]: I1008 00:21:52.886659 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/52664964-4de9-49b7-af7f-b97194774460-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "52664964-4de9-49b7-af7f-b97194774460" (UID: "52664964-4de9-49b7-af7f-b97194774460"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:21:52 crc kubenswrapper[4967]: I1008 00:21:52.973057 4967 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/52664964-4de9-49b7-af7f-b97194774460-build-blob-cache\") on node \"crc\" DevicePath \"\"" Oct 08 00:21:53 crc kubenswrapper[4967]: I1008 00:21:53.214636 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-2-build" event={"ID":"52664964-4de9-49b7-af7f-b97194774460","Type":"ContainerDied","Data":"04f5ec5005e7aa5c6dc2246bd3e8a13bcaddb21a468a8ff83c84cfa122411dc1"} Oct 08 00:21:53 crc kubenswrapper[4967]: I1008 00:21:53.214755 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="04f5ec5005e7aa5c6dc2246bd3e8a13bcaddb21a468a8ff83c84cfa122411dc1" Oct 08 00:21:53 crc kubenswrapper[4967]: I1008 00:21:53.214883 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-2-build" Oct 08 00:21:55 crc kubenswrapper[4967]: I1008 00:21:55.360560 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/52664964-4de9-49b7-af7f-b97194774460-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "52664964-4de9-49b7-af7f-b97194774460" (UID: "52664964-4de9-49b7-af7f-b97194774460"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:21:55 crc kubenswrapper[4967]: I1008 00:21:55.408857 4967 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/52664964-4de9-49b7-af7f-b97194774460-container-storage-root\") on node \"crc\" DevicePath \"\"" Oct 08 00:21:57 crc kubenswrapper[4967]: I1008 00:21:57.411888 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/sg-core-1-build"] Oct 08 00:21:57 crc kubenswrapper[4967]: E1008 00:21:57.412276 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52664964-4de9-49b7-af7f-b97194774460" containerName="docker-build" Oct 08 00:21:57 crc kubenswrapper[4967]: I1008 00:21:57.412300 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="52664964-4de9-49b7-af7f-b97194774460" containerName="docker-build" Oct 08 00:21:57 crc kubenswrapper[4967]: E1008 00:21:57.412325 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52664964-4de9-49b7-af7f-b97194774460" containerName="git-clone" Oct 08 00:21:57 crc kubenswrapper[4967]: I1008 00:21:57.412338 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="52664964-4de9-49b7-af7f-b97194774460" containerName="git-clone" Oct 08 00:21:57 crc kubenswrapper[4967]: E1008 00:21:57.412368 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f47a4120-aa3b-400b-aa4f-42b3262efd73" containerName="manage-dockerfile" Oct 08 00:21:57 crc kubenswrapper[4967]: I1008 00:21:57.412382 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="f47a4120-aa3b-400b-aa4f-42b3262efd73" containerName="manage-dockerfile" Oct 08 00:21:57 crc kubenswrapper[4967]: E1008 00:21:57.412403 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52664964-4de9-49b7-af7f-b97194774460" containerName="manage-dockerfile" Oct 08 00:21:57 crc kubenswrapper[4967]: I1008 00:21:57.412416 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="52664964-4de9-49b7-af7f-b97194774460" containerName="manage-dockerfile" Oct 08 00:21:57 crc kubenswrapper[4967]: E1008 00:21:57.412440 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f47a4120-aa3b-400b-aa4f-42b3262efd73" containerName="docker-build" Oct 08 00:21:57 crc kubenswrapper[4967]: I1008 00:21:57.412453 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="f47a4120-aa3b-400b-aa4f-42b3262efd73" containerName="docker-build" Oct 08 00:21:57 crc kubenswrapper[4967]: I1008 00:21:57.412636 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="f47a4120-aa3b-400b-aa4f-42b3262efd73" containerName="docker-build" Oct 08 00:21:57 crc kubenswrapper[4967]: I1008 00:21:57.412657 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="52664964-4de9-49b7-af7f-b97194774460" containerName="docker-build" Oct 08 00:21:57 crc kubenswrapper[4967]: I1008 00:21:57.414417 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-core-1-build" Oct 08 00:21:57 crc kubenswrapper[4967]: I1008 00:21:57.420516 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"sg-core-1-sys-config" Oct 08 00:21:57 crc kubenswrapper[4967]: I1008 00:21:57.420581 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"sg-core-1-ca" Oct 08 00:21:57 crc kubenswrapper[4967]: I1008 00:21:57.420817 4967 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-r7snr" Oct 08 00:21:57 crc kubenswrapper[4967]: I1008 00:21:57.420851 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"sg-core-1-global-ca" Oct 08 00:21:57 crc kubenswrapper[4967]: I1008 00:21:57.433580 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/sg-core-1-build"] Oct 08 00:21:57 crc kubenswrapper[4967]: I1008 00:21:57.540914 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/91ca438c-c884-44ec-b17b-69c0d6fd9195-build-system-configs\") pod \"sg-core-1-build\" (UID: \"91ca438c-c884-44ec-b17b-69c0d6fd9195\") " pod="service-telemetry/sg-core-1-build" Oct 08 00:21:57 crc kubenswrapper[4967]: I1008 00:21:57.540978 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/91ca438c-c884-44ec-b17b-69c0d6fd9195-build-blob-cache\") pod \"sg-core-1-build\" (UID: \"91ca438c-c884-44ec-b17b-69c0d6fd9195\") " pod="service-telemetry/sg-core-1-build" Oct 08 00:21:57 crc kubenswrapper[4967]: I1008 00:21:57.541013 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-r7snr-push\" (UniqueName: \"kubernetes.io/secret/91ca438c-c884-44ec-b17b-69c0d6fd9195-builder-dockercfg-r7snr-push\") pod \"sg-core-1-build\" (UID: \"91ca438c-c884-44ec-b17b-69c0d6fd9195\") " pod="service-telemetry/sg-core-1-build" Oct 08 00:21:57 crc kubenswrapper[4967]: I1008 00:21:57.541036 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/91ca438c-c884-44ec-b17b-69c0d6fd9195-build-proxy-ca-bundles\") pod \"sg-core-1-build\" (UID: \"91ca438c-c884-44ec-b17b-69c0d6fd9195\") " pod="service-telemetry/sg-core-1-build" Oct 08 00:21:57 crc kubenswrapper[4967]: I1008 00:21:57.541197 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/91ca438c-c884-44ec-b17b-69c0d6fd9195-build-ca-bundles\") pod \"sg-core-1-build\" (UID: \"91ca438c-c884-44ec-b17b-69c0d6fd9195\") " pod="service-telemetry/sg-core-1-build" Oct 08 00:21:57 crc kubenswrapper[4967]: I1008 00:21:57.541284 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/91ca438c-c884-44ec-b17b-69c0d6fd9195-buildcachedir\") pod \"sg-core-1-build\" (UID: \"91ca438c-c884-44ec-b17b-69c0d6fd9195\") " pod="service-telemetry/sg-core-1-build" Oct 08 00:21:57 crc kubenswrapper[4967]: I1008 00:21:57.541325 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-r7snr-pull\" (UniqueName: \"kubernetes.io/secret/91ca438c-c884-44ec-b17b-69c0d6fd9195-builder-dockercfg-r7snr-pull\") pod \"sg-core-1-build\" (UID: \"91ca438c-c884-44ec-b17b-69c0d6fd9195\") " pod="service-telemetry/sg-core-1-build" Oct 08 00:21:57 crc kubenswrapper[4967]: I1008 00:21:57.541447 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/91ca438c-c884-44ec-b17b-69c0d6fd9195-container-storage-run\") pod \"sg-core-1-build\" (UID: \"91ca438c-c884-44ec-b17b-69c0d6fd9195\") " pod="service-telemetry/sg-core-1-build" Oct 08 00:21:57 crc kubenswrapper[4967]: I1008 00:21:57.541478 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/91ca438c-c884-44ec-b17b-69c0d6fd9195-node-pullsecrets\") pod \"sg-core-1-build\" (UID: \"91ca438c-c884-44ec-b17b-69c0d6fd9195\") " pod="service-telemetry/sg-core-1-build" Oct 08 00:21:57 crc kubenswrapper[4967]: I1008 00:21:57.541502 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/91ca438c-c884-44ec-b17b-69c0d6fd9195-buildworkdir\") pod \"sg-core-1-build\" (UID: \"91ca438c-c884-44ec-b17b-69c0d6fd9195\") " pod="service-telemetry/sg-core-1-build" Oct 08 00:21:57 crc kubenswrapper[4967]: I1008 00:21:57.541525 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/91ca438c-c884-44ec-b17b-69c0d6fd9195-container-storage-root\") pod \"sg-core-1-build\" (UID: \"91ca438c-c884-44ec-b17b-69c0d6fd9195\") " pod="service-telemetry/sg-core-1-build" Oct 08 00:21:57 crc kubenswrapper[4967]: I1008 00:21:57.541549 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-868xn\" (UniqueName: \"kubernetes.io/projected/91ca438c-c884-44ec-b17b-69c0d6fd9195-kube-api-access-868xn\") pod \"sg-core-1-build\" (UID: \"91ca438c-c884-44ec-b17b-69c0d6fd9195\") " pod="service-telemetry/sg-core-1-build" Oct 08 00:21:57 crc kubenswrapper[4967]: I1008 00:21:57.642698 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/91ca438c-c884-44ec-b17b-69c0d6fd9195-container-storage-run\") pod \"sg-core-1-build\" (UID: \"91ca438c-c884-44ec-b17b-69c0d6fd9195\") " pod="service-telemetry/sg-core-1-build" Oct 08 00:21:57 crc kubenswrapper[4967]: I1008 00:21:57.642754 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/91ca438c-c884-44ec-b17b-69c0d6fd9195-node-pullsecrets\") pod \"sg-core-1-build\" (UID: \"91ca438c-c884-44ec-b17b-69c0d6fd9195\") " pod="service-telemetry/sg-core-1-build" Oct 08 00:21:57 crc kubenswrapper[4967]: I1008 00:21:57.642780 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/91ca438c-c884-44ec-b17b-69c0d6fd9195-buildworkdir\") pod \"sg-core-1-build\" (UID: \"91ca438c-c884-44ec-b17b-69c0d6fd9195\") " pod="service-telemetry/sg-core-1-build" Oct 08 00:21:57 crc kubenswrapper[4967]: I1008 00:21:57.642804 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/91ca438c-c884-44ec-b17b-69c0d6fd9195-container-storage-root\") pod \"sg-core-1-build\" (UID: \"91ca438c-c884-44ec-b17b-69c0d6fd9195\") " pod="service-telemetry/sg-core-1-build" Oct 08 00:21:57 crc kubenswrapper[4967]: I1008 00:21:57.642828 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-868xn\" (UniqueName: \"kubernetes.io/projected/91ca438c-c884-44ec-b17b-69c0d6fd9195-kube-api-access-868xn\") pod \"sg-core-1-build\" (UID: \"91ca438c-c884-44ec-b17b-69c0d6fd9195\") " pod="service-telemetry/sg-core-1-build" Oct 08 00:21:57 crc kubenswrapper[4967]: I1008 00:21:57.642870 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/91ca438c-c884-44ec-b17b-69c0d6fd9195-build-system-configs\") pod \"sg-core-1-build\" (UID: \"91ca438c-c884-44ec-b17b-69c0d6fd9195\") " pod="service-telemetry/sg-core-1-build" Oct 08 00:21:57 crc kubenswrapper[4967]: I1008 00:21:57.642894 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/91ca438c-c884-44ec-b17b-69c0d6fd9195-build-blob-cache\") pod \"sg-core-1-build\" (UID: \"91ca438c-c884-44ec-b17b-69c0d6fd9195\") " pod="service-telemetry/sg-core-1-build" Oct 08 00:21:57 crc kubenswrapper[4967]: I1008 00:21:57.642924 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-r7snr-push\" (UniqueName: \"kubernetes.io/secret/91ca438c-c884-44ec-b17b-69c0d6fd9195-builder-dockercfg-r7snr-push\") pod \"sg-core-1-build\" (UID: \"91ca438c-c884-44ec-b17b-69c0d6fd9195\") " pod="service-telemetry/sg-core-1-build" Oct 08 00:21:57 crc kubenswrapper[4967]: I1008 00:21:57.642953 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/91ca438c-c884-44ec-b17b-69c0d6fd9195-build-proxy-ca-bundles\") pod \"sg-core-1-build\" (UID: \"91ca438c-c884-44ec-b17b-69c0d6fd9195\") " pod="service-telemetry/sg-core-1-build" Oct 08 00:21:57 crc kubenswrapper[4967]: I1008 00:21:57.642980 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/91ca438c-c884-44ec-b17b-69c0d6fd9195-build-ca-bundles\") pod \"sg-core-1-build\" (UID: \"91ca438c-c884-44ec-b17b-69c0d6fd9195\") " pod="service-telemetry/sg-core-1-build" Oct 08 00:21:57 crc kubenswrapper[4967]: I1008 00:21:57.643008 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/91ca438c-c884-44ec-b17b-69c0d6fd9195-buildcachedir\") pod \"sg-core-1-build\" (UID: \"91ca438c-c884-44ec-b17b-69c0d6fd9195\") " pod="service-telemetry/sg-core-1-build" Oct 08 00:21:57 crc kubenswrapper[4967]: I1008 00:21:57.643031 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-r7snr-pull\" (UniqueName: \"kubernetes.io/secret/91ca438c-c884-44ec-b17b-69c0d6fd9195-builder-dockercfg-r7snr-pull\") pod \"sg-core-1-build\" (UID: \"91ca438c-c884-44ec-b17b-69c0d6fd9195\") " pod="service-telemetry/sg-core-1-build" Oct 08 00:21:57 crc kubenswrapper[4967]: I1008 00:21:57.642952 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/91ca438c-c884-44ec-b17b-69c0d6fd9195-node-pullsecrets\") pod \"sg-core-1-build\" (UID: \"91ca438c-c884-44ec-b17b-69c0d6fd9195\") " pod="service-telemetry/sg-core-1-build" Oct 08 00:21:57 crc kubenswrapper[4967]: I1008 00:21:57.643293 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/91ca438c-c884-44ec-b17b-69c0d6fd9195-buildcachedir\") pod \"sg-core-1-build\" (UID: \"91ca438c-c884-44ec-b17b-69c0d6fd9195\") " pod="service-telemetry/sg-core-1-build" Oct 08 00:21:57 crc kubenswrapper[4967]: I1008 00:21:57.643495 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/91ca438c-c884-44ec-b17b-69c0d6fd9195-buildworkdir\") pod \"sg-core-1-build\" (UID: \"91ca438c-c884-44ec-b17b-69c0d6fd9195\") " pod="service-telemetry/sg-core-1-build" Oct 08 00:21:57 crc kubenswrapper[4967]: I1008 00:21:57.643819 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/91ca438c-c884-44ec-b17b-69c0d6fd9195-container-storage-root\") pod \"sg-core-1-build\" (UID: \"91ca438c-c884-44ec-b17b-69c0d6fd9195\") " pod="service-telemetry/sg-core-1-build" Oct 08 00:21:57 crc kubenswrapper[4967]: I1008 00:21:57.643840 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/91ca438c-c884-44ec-b17b-69c0d6fd9195-container-storage-run\") pod \"sg-core-1-build\" (UID: \"91ca438c-c884-44ec-b17b-69c0d6fd9195\") " pod="service-telemetry/sg-core-1-build" Oct 08 00:21:57 crc kubenswrapper[4967]: I1008 00:21:57.643969 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/91ca438c-c884-44ec-b17b-69c0d6fd9195-build-blob-cache\") pod \"sg-core-1-build\" (UID: \"91ca438c-c884-44ec-b17b-69c0d6fd9195\") " pod="service-telemetry/sg-core-1-build" Oct 08 00:21:57 crc kubenswrapper[4967]: I1008 00:21:57.644376 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/91ca438c-c884-44ec-b17b-69c0d6fd9195-build-ca-bundles\") pod \"sg-core-1-build\" (UID: \"91ca438c-c884-44ec-b17b-69c0d6fd9195\") " pod="service-telemetry/sg-core-1-build" Oct 08 00:21:57 crc kubenswrapper[4967]: I1008 00:21:57.644439 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/91ca438c-c884-44ec-b17b-69c0d6fd9195-build-proxy-ca-bundles\") pod \"sg-core-1-build\" (UID: \"91ca438c-c884-44ec-b17b-69c0d6fd9195\") " pod="service-telemetry/sg-core-1-build" Oct 08 00:21:57 crc kubenswrapper[4967]: I1008 00:21:57.645482 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/91ca438c-c884-44ec-b17b-69c0d6fd9195-build-system-configs\") pod \"sg-core-1-build\" (UID: \"91ca438c-c884-44ec-b17b-69c0d6fd9195\") " pod="service-telemetry/sg-core-1-build" Oct 08 00:21:57 crc kubenswrapper[4967]: I1008 00:21:57.648338 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-r7snr-pull\" (UniqueName: \"kubernetes.io/secret/91ca438c-c884-44ec-b17b-69c0d6fd9195-builder-dockercfg-r7snr-pull\") pod \"sg-core-1-build\" (UID: \"91ca438c-c884-44ec-b17b-69c0d6fd9195\") " pod="service-telemetry/sg-core-1-build" Oct 08 00:21:57 crc kubenswrapper[4967]: I1008 00:21:57.650090 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-r7snr-push\" (UniqueName: \"kubernetes.io/secret/91ca438c-c884-44ec-b17b-69c0d6fd9195-builder-dockercfg-r7snr-push\") pod \"sg-core-1-build\" (UID: \"91ca438c-c884-44ec-b17b-69c0d6fd9195\") " pod="service-telemetry/sg-core-1-build" Oct 08 00:21:57 crc kubenswrapper[4967]: I1008 00:21:57.667086 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-868xn\" (UniqueName: \"kubernetes.io/projected/91ca438c-c884-44ec-b17b-69c0d6fd9195-kube-api-access-868xn\") pod \"sg-core-1-build\" (UID: \"91ca438c-c884-44ec-b17b-69c0d6fd9195\") " pod="service-telemetry/sg-core-1-build" Oct 08 00:21:57 crc kubenswrapper[4967]: I1008 00:21:57.742041 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-core-1-build" Oct 08 00:21:58 crc kubenswrapper[4967]: I1008 00:21:58.225838 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/sg-core-1-build"] Oct 08 00:21:58 crc kubenswrapper[4967]: I1008 00:21:58.254803 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-1-build" event={"ID":"91ca438c-c884-44ec-b17b-69c0d6fd9195","Type":"ContainerStarted","Data":"5dc6602deba119605c9b32ef5e92e28f3eb06d485e11739f5293e6928f2e021a"} Oct 08 00:21:59 crc kubenswrapper[4967]: I1008 00:21:59.265637 4967 generic.go:334] "Generic (PLEG): container finished" podID="91ca438c-c884-44ec-b17b-69c0d6fd9195" containerID="cd760195179f9866f81f101dd745f3c0e447d1508586071d6c90d008791746b5" exitCode=0 Oct 08 00:21:59 crc kubenswrapper[4967]: I1008 00:21:59.265737 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-1-build" event={"ID":"91ca438c-c884-44ec-b17b-69c0d6fd9195","Type":"ContainerDied","Data":"cd760195179f9866f81f101dd745f3c0e447d1508586071d6c90d008791746b5"} Oct 08 00:22:00 crc kubenswrapper[4967]: I1008 00:22:00.276631 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-1-build" event={"ID":"91ca438c-c884-44ec-b17b-69c0d6fd9195","Type":"ContainerStarted","Data":"e1562aa1d2e7281f35c008de9afa7e5d7c94c7a8d30b07c16bc644d40d05429c"} Oct 08 00:22:00 crc kubenswrapper[4967]: I1008 00:22:00.310829 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/sg-core-1-build" podStartSLOduration=3.310805526 podStartE2EDuration="3.310805526s" podCreationTimestamp="2025-10-08 00:21:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:22:00.306950539 +0000 UTC m=+918.957485238" watchObservedRunningTime="2025-10-08 00:22:00.310805526 +0000 UTC m=+918.961340215" Oct 08 00:22:07 crc kubenswrapper[4967]: I1008 00:22:07.612744 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/sg-core-1-build"] Oct 08 00:22:07 crc kubenswrapper[4967]: I1008 00:22:07.613530 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="service-telemetry/sg-core-1-build" podUID="91ca438c-c884-44ec-b17b-69c0d6fd9195" containerName="docker-build" containerID="cri-o://e1562aa1d2e7281f35c008de9afa7e5d7c94c7a8d30b07c16bc644d40d05429c" gracePeriod=30 Oct 08 00:22:08 crc kubenswrapper[4967]: I1008 00:22:08.008291 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_sg-core-1-build_91ca438c-c884-44ec-b17b-69c0d6fd9195/docker-build/0.log" Oct 08 00:22:08 crc kubenswrapper[4967]: I1008 00:22:08.008842 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-core-1-build" Oct 08 00:22:08 crc kubenswrapper[4967]: I1008 00:22:08.021042 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/91ca438c-c884-44ec-b17b-69c0d6fd9195-build-proxy-ca-bundles\") pod \"91ca438c-c884-44ec-b17b-69c0d6fd9195\" (UID: \"91ca438c-c884-44ec-b17b-69c0d6fd9195\") " Oct 08 00:22:08 crc kubenswrapper[4967]: I1008 00:22:08.021114 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-868xn\" (UniqueName: \"kubernetes.io/projected/91ca438c-c884-44ec-b17b-69c0d6fd9195-kube-api-access-868xn\") pod \"91ca438c-c884-44ec-b17b-69c0d6fd9195\" (UID: \"91ca438c-c884-44ec-b17b-69c0d6fd9195\") " Oct 08 00:22:08 crc kubenswrapper[4967]: I1008 00:22:08.021144 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/91ca438c-c884-44ec-b17b-69c0d6fd9195-build-blob-cache\") pod \"91ca438c-c884-44ec-b17b-69c0d6fd9195\" (UID: \"91ca438c-c884-44ec-b17b-69c0d6fd9195\") " Oct 08 00:22:08 crc kubenswrapper[4967]: I1008 00:22:08.021177 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/91ca438c-c884-44ec-b17b-69c0d6fd9195-container-storage-root\") pod \"91ca438c-c884-44ec-b17b-69c0d6fd9195\" (UID: \"91ca438c-c884-44ec-b17b-69c0d6fd9195\") " Oct 08 00:22:08 crc kubenswrapper[4967]: I1008 00:22:08.021206 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-r7snr-push\" (UniqueName: \"kubernetes.io/secret/91ca438c-c884-44ec-b17b-69c0d6fd9195-builder-dockercfg-r7snr-push\") pod \"91ca438c-c884-44ec-b17b-69c0d6fd9195\" (UID: \"91ca438c-c884-44ec-b17b-69c0d6fd9195\") " Oct 08 00:22:08 crc kubenswrapper[4967]: I1008 00:22:08.021226 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/91ca438c-c884-44ec-b17b-69c0d6fd9195-buildworkdir\") pod \"91ca438c-c884-44ec-b17b-69c0d6fd9195\" (UID: \"91ca438c-c884-44ec-b17b-69c0d6fd9195\") " Oct 08 00:22:08 crc kubenswrapper[4967]: I1008 00:22:08.021280 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/91ca438c-c884-44ec-b17b-69c0d6fd9195-container-storage-run\") pod \"91ca438c-c884-44ec-b17b-69c0d6fd9195\" (UID: \"91ca438c-c884-44ec-b17b-69c0d6fd9195\") " Oct 08 00:22:08 crc kubenswrapper[4967]: I1008 00:22:08.021312 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/91ca438c-c884-44ec-b17b-69c0d6fd9195-node-pullsecrets\") pod \"91ca438c-c884-44ec-b17b-69c0d6fd9195\" (UID: \"91ca438c-c884-44ec-b17b-69c0d6fd9195\") " Oct 08 00:22:08 crc kubenswrapper[4967]: I1008 00:22:08.021338 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/91ca438c-c884-44ec-b17b-69c0d6fd9195-build-ca-bundles\") pod \"91ca438c-c884-44ec-b17b-69c0d6fd9195\" (UID: \"91ca438c-c884-44ec-b17b-69c0d6fd9195\") " Oct 08 00:22:08 crc kubenswrapper[4967]: I1008 00:22:08.021359 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/91ca438c-c884-44ec-b17b-69c0d6fd9195-buildcachedir\") pod \"91ca438c-c884-44ec-b17b-69c0d6fd9195\" (UID: \"91ca438c-c884-44ec-b17b-69c0d6fd9195\") " Oct 08 00:22:08 crc kubenswrapper[4967]: I1008 00:22:08.021394 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/91ca438c-c884-44ec-b17b-69c0d6fd9195-build-system-configs\") pod \"91ca438c-c884-44ec-b17b-69c0d6fd9195\" (UID: \"91ca438c-c884-44ec-b17b-69c0d6fd9195\") " Oct 08 00:22:08 crc kubenswrapper[4967]: I1008 00:22:08.021418 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-r7snr-pull\" (UniqueName: \"kubernetes.io/secret/91ca438c-c884-44ec-b17b-69c0d6fd9195-builder-dockercfg-r7snr-pull\") pod \"91ca438c-c884-44ec-b17b-69c0d6fd9195\" (UID: \"91ca438c-c884-44ec-b17b-69c0d6fd9195\") " Oct 08 00:22:08 crc kubenswrapper[4967]: I1008 00:22:08.021702 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/91ca438c-c884-44ec-b17b-69c0d6fd9195-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "91ca438c-c884-44ec-b17b-69c0d6fd9195" (UID: "91ca438c-c884-44ec-b17b-69c0d6fd9195"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:22:08 crc kubenswrapper[4967]: I1008 00:22:08.022029 4967 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/91ca438c-c884-44ec-b17b-69c0d6fd9195-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 08 00:22:08 crc kubenswrapper[4967]: I1008 00:22:08.022054 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91ca438c-c884-44ec-b17b-69c0d6fd9195-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "91ca438c-c884-44ec-b17b-69c0d6fd9195" (UID: "91ca438c-c884-44ec-b17b-69c0d6fd9195"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:22:08 crc kubenswrapper[4967]: I1008 00:22:08.022061 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/91ca438c-c884-44ec-b17b-69c0d6fd9195-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "91ca438c-c884-44ec-b17b-69c0d6fd9195" (UID: "91ca438c-c884-44ec-b17b-69c0d6fd9195"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 00:22:08 crc kubenswrapper[4967]: I1008 00:22:08.022084 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/91ca438c-c884-44ec-b17b-69c0d6fd9195-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "91ca438c-c884-44ec-b17b-69c0d6fd9195" (UID: "91ca438c-c884-44ec-b17b-69c0d6fd9195"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 00:22:08 crc kubenswrapper[4967]: I1008 00:22:08.022942 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/91ca438c-c884-44ec-b17b-69c0d6fd9195-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "91ca438c-c884-44ec-b17b-69c0d6fd9195" (UID: "91ca438c-c884-44ec-b17b-69c0d6fd9195"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:22:08 crc kubenswrapper[4967]: I1008 00:22:08.023761 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/91ca438c-c884-44ec-b17b-69c0d6fd9195-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "91ca438c-c884-44ec-b17b-69c0d6fd9195" (UID: "91ca438c-c884-44ec-b17b-69c0d6fd9195"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:22:08 crc kubenswrapper[4967]: I1008 00:22:08.023981 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91ca438c-c884-44ec-b17b-69c0d6fd9195-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "91ca438c-c884-44ec-b17b-69c0d6fd9195" (UID: "91ca438c-c884-44ec-b17b-69c0d6fd9195"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:22:08 crc kubenswrapper[4967]: I1008 00:22:08.028636 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91ca438c-c884-44ec-b17b-69c0d6fd9195-kube-api-access-868xn" (OuterVolumeSpecName: "kube-api-access-868xn") pod "91ca438c-c884-44ec-b17b-69c0d6fd9195" (UID: "91ca438c-c884-44ec-b17b-69c0d6fd9195"). InnerVolumeSpecName "kube-api-access-868xn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:22:08 crc kubenswrapper[4967]: I1008 00:22:08.032224 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91ca438c-c884-44ec-b17b-69c0d6fd9195-builder-dockercfg-r7snr-pull" (OuterVolumeSpecName: "builder-dockercfg-r7snr-pull") pod "91ca438c-c884-44ec-b17b-69c0d6fd9195" (UID: "91ca438c-c884-44ec-b17b-69c0d6fd9195"). InnerVolumeSpecName "builder-dockercfg-r7snr-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:22:08 crc kubenswrapper[4967]: I1008 00:22:08.033225 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91ca438c-c884-44ec-b17b-69c0d6fd9195-builder-dockercfg-r7snr-push" (OuterVolumeSpecName: "builder-dockercfg-r7snr-push") pod "91ca438c-c884-44ec-b17b-69c0d6fd9195" (UID: "91ca438c-c884-44ec-b17b-69c0d6fd9195"). InnerVolumeSpecName "builder-dockercfg-r7snr-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:22:08 crc kubenswrapper[4967]: I1008 00:22:08.123198 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-868xn\" (UniqueName: \"kubernetes.io/projected/91ca438c-c884-44ec-b17b-69c0d6fd9195-kube-api-access-868xn\") on node \"crc\" DevicePath \"\"" Oct 08 00:22:08 crc kubenswrapper[4967]: I1008 00:22:08.123265 4967 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-r7snr-push\" (UniqueName: \"kubernetes.io/secret/91ca438c-c884-44ec-b17b-69c0d6fd9195-builder-dockercfg-r7snr-push\") on node \"crc\" DevicePath \"\"" Oct 08 00:22:08 crc kubenswrapper[4967]: I1008 00:22:08.123279 4967 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/91ca438c-c884-44ec-b17b-69c0d6fd9195-buildworkdir\") on node \"crc\" DevicePath \"\"" Oct 08 00:22:08 crc kubenswrapper[4967]: I1008 00:22:08.123290 4967 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/91ca438c-c884-44ec-b17b-69c0d6fd9195-container-storage-run\") on node \"crc\" DevicePath \"\"" Oct 08 00:22:08 crc kubenswrapper[4967]: I1008 00:22:08.123302 4967 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/91ca438c-c884-44ec-b17b-69c0d6fd9195-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Oct 08 00:22:08 crc kubenswrapper[4967]: I1008 00:22:08.123312 4967 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/91ca438c-c884-44ec-b17b-69c0d6fd9195-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 08 00:22:08 crc kubenswrapper[4967]: I1008 00:22:08.123321 4967 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/91ca438c-c884-44ec-b17b-69c0d6fd9195-buildcachedir\") on node \"crc\" DevicePath \"\"" Oct 08 00:22:08 crc kubenswrapper[4967]: I1008 00:22:08.123331 4967 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/91ca438c-c884-44ec-b17b-69c0d6fd9195-build-system-configs\") on node \"crc\" DevicePath \"\"" Oct 08 00:22:08 crc kubenswrapper[4967]: I1008 00:22:08.123342 4967 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-r7snr-pull\" (UniqueName: \"kubernetes.io/secret/91ca438c-c884-44ec-b17b-69c0d6fd9195-builder-dockercfg-r7snr-pull\") on node \"crc\" DevicePath \"\"" Oct 08 00:22:08 crc kubenswrapper[4967]: I1008 00:22:08.124371 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91ca438c-c884-44ec-b17b-69c0d6fd9195-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "91ca438c-c884-44ec-b17b-69c0d6fd9195" (UID: "91ca438c-c884-44ec-b17b-69c0d6fd9195"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:22:08 crc kubenswrapper[4967]: I1008 00:22:08.181847 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91ca438c-c884-44ec-b17b-69c0d6fd9195-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "91ca438c-c884-44ec-b17b-69c0d6fd9195" (UID: "91ca438c-c884-44ec-b17b-69c0d6fd9195"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:22:08 crc kubenswrapper[4967]: I1008 00:22:08.224432 4967 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/91ca438c-c884-44ec-b17b-69c0d6fd9195-build-blob-cache\") on node \"crc\" DevicePath \"\"" Oct 08 00:22:08 crc kubenswrapper[4967]: I1008 00:22:08.224481 4967 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/91ca438c-c884-44ec-b17b-69c0d6fd9195-container-storage-root\") on node \"crc\" DevicePath \"\"" Oct 08 00:22:08 crc kubenswrapper[4967]: I1008 00:22:08.332348 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_sg-core-1-build_91ca438c-c884-44ec-b17b-69c0d6fd9195/docker-build/0.log" Oct 08 00:22:08 crc kubenswrapper[4967]: I1008 00:22:08.333291 4967 generic.go:334] "Generic (PLEG): container finished" podID="91ca438c-c884-44ec-b17b-69c0d6fd9195" containerID="e1562aa1d2e7281f35c008de9afa7e5d7c94c7a8d30b07c16bc644d40d05429c" exitCode=1 Oct 08 00:22:08 crc kubenswrapper[4967]: I1008 00:22:08.333343 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-1-build" event={"ID":"91ca438c-c884-44ec-b17b-69c0d6fd9195","Type":"ContainerDied","Data":"e1562aa1d2e7281f35c008de9afa7e5d7c94c7a8d30b07c16bc644d40d05429c"} Oct 08 00:22:08 crc kubenswrapper[4967]: I1008 00:22:08.333372 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-1-build" event={"ID":"91ca438c-c884-44ec-b17b-69c0d6fd9195","Type":"ContainerDied","Data":"5dc6602deba119605c9b32ef5e92e28f3eb06d485e11739f5293e6928f2e021a"} Oct 08 00:22:08 crc kubenswrapper[4967]: I1008 00:22:08.333388 4967 scope.go:117] "RemoveContainer" containerID="e1562aa1d2e7281f35c008de9afa7e5d7c94c7a8d30b07c16bc644d40d05429c" Oct 08 00:22:08 crc kubenswrapper[4967]: I1008 00:22:08.333416 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-core-1-build" Oct 08 00:22:08 crc kubenswrapper[4967]: I1008 00:22:08.377348 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/sg-core-1-build"] Oct 08 00:22:08 crc kubenswrapper[4967]: I1008 00:22:08.384890 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["service-telemetry/sg-core-1-build"] Oct 08 00:22:08 crc kubenswrapper[4967]: I1008 00:22:08.409563 4967 scope.go:117] "RemoveContainer" containerID="cd760195179f9866f81f101dd745f3c0e447d1508586071d6c90d008791746b5" Oct 08 00:22:08 crc kubenswrapper[4967]: I1008 00:22:08.435317 4967 scope.go:117] "RemoveContainer" containerID="e1562aa1d2e7281f35c008de9afa7e5d7c94c7a8d30b07c16bc644d40d05429c" Oct 08 00:22:08 crc kubenswrapper[4967]: E1008 00:22:08.435737 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e1562aa1d2e7281f35c008de9afa7e5d7c94c7a8d30b07c16bc644d40d05429c\": container with ID starting with e1562aa1d2e7281f35c008de9afa7e5d7c94c7a8d30b07c16bc644d40d05429c not found: ID does not exist" containerID="e1562aa1d2e7281f35c008de9afa7e5d7c94c7a8d30b07c16bc644d40d05429c" Oct 08 00:22:08 crc kubenswrapper[4967]: I1008 00:22:08.435777 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1562aa1d2e7281f35c008de9afa7e5d7c94c7a8d30b07c16bc644d40d05429c"} err="failed to get container status \"e1562aa1d2e7281f35c008de9afa7e5d7c94c7a8d30b07c16bc644d40d05429c\": rpc error: code = NotFound desc = could not find container \"e1562aa1d2e7281f35c008de9afa7e5d7c94c7a8d30b07c16bc644d40d05429c\": container with ID starting with e1562aa1d2e7281f35c008de9afa7e5d7c94c7a8d30b07c16bc644d40d05429c not found: ID does not exist" Oct 08 00:22:08 crc kubenswrapper[4967]: I1008 00:22:08.435804 4967 scope.go:117] "RemoveContainer" containerID="cd760195179f9866f81f101dd745f3c0e447d1508586071d6c90d008791746b5" Oct 08 00:22:08 crc kubenswrapper[4967]: E1008 00:22:08.437524 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd760195179f9866f81f101dd745f3c0e447d1508586071d6c90d008791746b5\": container with ID starting with cd760195179f9866f81f101dd745f3c0e447d1508586071d6c90d008791746b5 not found: ID does not exist" containerID="cd760195179f9866f81f101dd745f3c0e447d1508586071d6c90d008791746b5" Oct 08 00:22:08 crc kubenswrapper[4967]: I1008 00:22:08.437552 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd760195179f9866f81f101dd745f3c0e447d1508586071d6c90d008791746b5"} err="failed to get container status \"cd760195179f9866f81f101dd745f3c0e447d1508586071d6c90d008791746b5\": rpc error: code = NotFound desc = could not find container \"cd760195179f9866f81f101dd745f3c0e447d1508586071d6c90d008791746b5\": container with ID starting with cd760195179f9866f81f101dd745f3c0e447d1508586071d6c90d008791746b5 not found: ID does not exist" Oct 08 00:22:09 crc kubenswrapper[4967]: I1008 00:22:09.253050 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/sg-core-2-build"] Oct 08 00:22:09 crc kubenswrapper[4967]: E1008 00:22:09.253503 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91ca438c-c884-44ec-b17b-69c0d6fd9195" containerName="docker-build" Oct 08 00:22:09 crc kubenswrapper[4967]: I1008 00:22:09.253542 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="91ca438c-c884-44ec-b17b-69c0d6fd9195" containerName="docker-build" Oct 08 00:22:09 crc kubenswrapper[4967]: E1008 00:22:09.253587 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91ca438c-c884-44ec-b17b-69c0d6fd9195" containerName="manage-dockerfile" Oct 08 00:22:09 crc kubenswrapper[4967]: I1008 00:22:09.253606 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="91ca438c-c884-44ec-b17b-69c0d6fd9195" containerName="manage-dockerfile" Oct 08 00:22:09 crc kubenswrapper[4967]: I1008 00:22:09.253901 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="91ca438c-c884-44ec-b17b-69c0d6fd9195" containerName="docker-build" Oct 08 00:22:09 crc kubenswrapper[4967]: I1008 00:22:09.256009 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-core-2-build" Oct 08 00:22:09 crc kubenswrapper[4967]: I1008 00:22:09.260424 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"sg-core-2-global-ca" Oct 08 00:22:09 crc kubenswrapper[4967]: I1008 00:22:09.260492 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"sg-core-2-sys-config" Oct 08 00:22:09 crc kubenswrapper[4967]: I1008 00:22:09.261889 4967 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-r7snr" Oct 08 00:22:09 crc kubenswrapper[4967]: I1008 00:22:09.263817 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"sg-core-2-ca" Oct 08 00:22:09 crc kubenswrapper[4967]: I1008 00:22:09.285353 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/sg-core-2-build"] Oct 08 00:22:09 crc kubenswrapper[4967]: I1008 00:22:09.341531 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-r7snr-pull\" (UniqueName: \"kubernetes.io/secret/6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc-builder-dockercfg-r7snr-pull\") pod \"sg-core-2-build\" (UID: \"6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc\") " pod="service-telemetry/sg-core-2-build" Oct 08 00:22:09 crc kubenswrapper[4967]: I1008 00:22:09.341581 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc-node-pullsecrets\") pod \"sg-core-2-build\" (UID: \"6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc\") " pod="service-telemetry/sg-core-2-build" Oct 08 00:22:09 crc kubenswrapper[4967]: I1008 00:22:09.341660 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc-build-ca-bundles\") pod \"sg-core-2-build\" (UID: \"6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc\") " pod="service-telemetry/sg-core-2-build" Oct 08 00:22:09 crc kubenswrapper[4967]: I1008 00:22:09.341745 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6b88q\" (UniqueName: \"kubernetes.io/projected/6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc-kube-api-access-6b88q\") pod \"sg-core-2-build\" (UID: \"6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc\") " pod="service-telemetry/sg-core-2-build" Oct 08 00:22:09 crc kubenswrapper[4967]: I1008 00:22:09.341781 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc-buildcachedir\") pod \"sg-core-2-build\" (UID: \"6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc\") " pod="service-telemetry/sg-core-2-build" Oct 08 00:22:09 crc kubenswrapper[4967]: I1008 00:22:09.341830 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-r7snr-push\" (UniqueName: \"kubernetes.io/secret/6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc-builder-dockercfg-r7snr-push\") pod \"sg-core-2-build\" (UID: \"6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc\") " pod="service-telemetry/sg-core-2-build" Oct 08 00:22:09 crc kubenswrapper[4967]: I1008 00:22:09.341878 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc-build-proxy-ca-bundles\") pod \"sg-core-2-build\" (UID: \"6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc\") " pod="service-telemetry/sg-core-2-build" Oct 08 00:22:09 crc kubenswrapper[4967]: I1008 00:22:09.341930 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc-container-storage-root\") pod \"sg-core-2-build\" (UID: \"6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc\") " pod="service-telemetry/sg-core-2-build" Oct 08 00:22:09 crc kubenswrapper[4967]: I1008 00:22:09.341998 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc-container-storage-run\") pod \"sg-core-2-build\" (UID: \"6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc\") " pod="service-telemetry/sg-core-2-build" Oct 08 00:22:09 crc kubenswrapper[4967]: I1008 00:22:09.342029 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc-build-blob-cache\") pod \"sg-core-2-build\" (UID: \"6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc\") " pod="service-telemetry/sg-core-2-build" Oct 08 00:22:09 crc kubenswrapper[4967]: I1008 00:22:09.342064 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc-build-system-configs\") pod \"sg-core-2-build\" (UID: \"6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc\") " pod="service-telemetry/sg-core-2-build" Oct 08 00:22:09 crc kubenswrapper[4967]: I1008 00:22:09.342099 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc-buildworkdir\") pod \"sg-core-2-build\" (UID: \"6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc\") " pod="service-telemetry/sg-core-2-build" Oct 08 00:22:09 crc kubenswrapper[4967]: I1008 00:22:09.443256 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc-build-proxy-ca-bundles\") pod \"sg-core-2-build\" (UID: \"6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc\") " pod="service-telemetry/sg-core-2-build" Oct 08 00:22:09 crc kubenswrapper[4967]: I1008 00:22:09.443312 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc-container-storage-root\") pod \"sg-core-2-build\" (UID: \"6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc\") " pod="service-telemetry/sg-core-2-build" Oct 08 00:22:09 crc kubenswrapper[4967]: I1008 00:22:09.443336 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc-container-storage-run\") pod \"sg-core-2-build\" (UID: \"6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc\") " pod="service-telemetry/sg-core-2-build" Oct 08 00:22:09 crc kubenswrapper[4967]: I1008 00:22:09.443353 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc-build-blob-cache\") pod \"sg-core-2-build\" (UID: \"6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc\") " pod="service-telemetry/sg-core-2-build" Oct 08 00:22:09 crc kubenswrapper[4967]: I1008 00:22:09.443372 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc-build-system-configs\") pod \"sg-core-2-build\" (UID: \"6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc\") " pod="service-telemetry/sg-core-2-build" Oct 08 00:22:09 crc kubenswrapper[4967]: I1008 00:22:09.443390 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc-buildworkdir\") pod \"sg-core-2-build\" (UID: \"6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc\") " pod="service-telemetry/sg-core-2-build" Oct 08 00:22:09 crc kubenswrapper[4967]: I1008 00:22:09.443422 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-r7snr-pull\" (UniqueName: \"kubernetes.io/secret/6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc-builder-dockercfg-r7snr-pull\") pod \"sg-core-2-build\" (UID: \"6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc\") " pod="service-telemetry/sg-core-2-build" Oct 08 00:22:09 crc kubenswrapper[4967]: I1008 00:22:09.443438 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc-node-pullsecrets\") pod \"sg-core-2-build\" (UID: \"6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc\") " pod="service-telemetry/sg-core-2-build" Oct 08 00:22:09 crc kubenswrapper[4967]: I1008 00:22:09.443459 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc-build-ca-bundles\") pod \"sg-core-2-build\" (UID: \"6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc\") " pod="service-telemetry/sg-core-2-build" Oct 08 00:22:09 crc kubenswrapper[4967]: I1008 00:22:09.443480 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6b88q\" (UniqueName: \"kubernetes.io/projected/6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc-kube-api-access-6b88q\") pod \"sg-core-2-build\" (UID: \"6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc\") " pod="service-telemetry/sg-core-2-build" Oct 08 00:22:09 crc kubenswrapper[4967]: I1008 00:22:09.443499 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc-buildcachedir\") pod \"sg-core-2-build\" (UID: \"6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc\") " pod="service-telemetry/sg-core-2-build" Oct 08 00:22:09 crc kubenswrapper[4967]: I1008 00:22:09.443513 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-r7snr-push\" (UniqueName: \"kubernetes.io/secret/6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc-builder-dockercfg-r7snr-push\") pod \"sg-core-2-build\" (UID: \"6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc\") " pod="service-telemetry/sg-core-2-build" Oct 08 00:22:09 crc kubenswrapper[4967]: I1008 00:22:09.444657 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc-buildworkdir\") pod \"sg-core-2-build\" (UID: \"6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc\") " pod="service-telemetry/sg-core-2-build" Oct 08 00:22:09 crc kubenswrapper[4967]: I1008 00:22:09.444828 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc-build-ca-bundles\") pod \"sg-core-2-build\" (UID: \"6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc\") " pod="service-telemetry/sg-core-2-build" Oct 08 00:22:09 crc kubenswrapper[4967]: I1008 00:22:09.444885 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc-node-pullsecrets\") pod \"sg-core-2-build\" (UID: \"6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc\") " pod="service-telemetry/sg-core-2-build" Oct 08 00:22:09 crc kubenswrapper[4967]: I1008 00:22:09.444919 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc-buildcachedir\") pod \"sg-core-2-build\" (UID: \"6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc\") " pod="service-telemetry/sg-core-2-build" Oct 08 00:22:09 crc kubenswrapper[4967]: I1008 00:22:09.445093 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc-container-storage-run\") pod \"sg-core-2-build\" (UID: \"6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc\") " pod="service-telemetry/sg-core-2-build" Oct 08 00:22:09 crc kubenswrapper[4967]: I1008 00:22:09.445303 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc-container-storage-root\") pod \"sg-core-2-build\" (UID: \"6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc\") " pod="service-telemetry/sg-core-2-build" Oct 08 00:22:09 crc kubenswrapper[4967]: I1008 00:22:09.445504 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc-build-blob-cache\") pod \"sg-core-2-build\" (UID: \"6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc\") " pod="service-telemetry/sg-core-2-build" Oct 08 00:22:09 crc kubenswrapper[4967]: I1008 00:22:09.445828 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc-build-proxy-ca-bundles\") pod \"sg-core-2-build\" (UID: \"6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc\") " pod="service-telemetry/sg-core-2-build" Oct 08 00:22:09 crc kubenswrapper[4967]: I1008 00:22:09.445879 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc-build-system-configs\") pod \"sg-core-2-build\" (UID: \"6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc\") " pod="service-telemetry/sg-core-2-build" Oct 08 00:22:09 crc kubenswrapper[4967]: I1008 00:22:09.448116 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-r7snr-push\" (UniqueName: \"kubernetes.io/secret/6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc-builder-dockercfg-r7snr-push\") pod \"sg-core-2-build\" (UID: \"6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc\") " pod="service-telemetry/sg-core-2-build" Oct 08 00:22:09 crc kubenswrapper[4967]: I1008 00:22:09.449396 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-r7snr-pull\" (UniqueName: \"kubernetes.io/secret/6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc-builder-dockercfg-r7snr-pull\") pod \"sg-core-2-build\" (UID: \"6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc\") " pod="service-telemetry/sg-core-2-build" Oct 08 00:22:09 crc kubenswrapper[4967]: I1008 00:22:09.461029 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6b88q\" (UniqueName: \"kubernetes.io/projected/6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc-kube-api-access-6b88q\") pod \"sg-core-2-build\" (UID: \"6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc\") " pod="service-telemetry/sg-core-2-build" Oct 08 00:22:09 crc kubenswrapper[4967]: I1008 00:22:09.592609 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-core-2-build" Oct 08 00:22:09 crc kubenswrapper[4967]: I1008 00:22:09.862256 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/sg-core-2-build"] Oct 08 00:22:09 crc kubenswrapper[4967]: W1008 00:22:09.871988 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6d02c04e_3a4d_405c_8dca_c9f7aaeb82bc.slice/crio-6334261af191189aa20688c2e6e805f2551dcab6688c2d544163a650ff8ae299 WatchSource:0}: Error finding container 6334261af191189aa20688c2e6e805f2551dcab6688c2d544163a650ff8ae299: Status 404 returned error can't find the container with id 6334261af191189aa20688c2e6e805f2551dcab6688c2d544163a650ff8ae299 Oct 08 00:22:10 crc kubenswrapper[4967]: I1008 00:22:10.125705 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="91ca438c-c884-44ec-b17b-69c0d6fd9195" path="/var/lib/kubelet/pods/91ca438c-c884-44ec-b17b-69c0d6fd9195/volumes" Oct 08 00:22:10 crc kubenswrapper[4967]: I1008 00:22:10.350008 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-2-build" event={"ID":"6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc","Type":"ContainerStarted","Data":"1215852094aff592dcbf374769e6d71692e8f6e30c02cab5c55382c5d14210a1"} Oct 08 00:22:10 crc kubenswrapper[4967]: I1008 00:22:10.350060 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-2-build" event={"ID":"6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc","Type":"ContainerStarted","Data":"6334261af191189aa20688c2e6e805f2551dcab6688c2d544163a650ff8ae299"} Oct 08 00:22:11 crc kubenswrapper[4967]: I1008 00:22:11.357816 4967 generic.go:334] "Generic (PLEG): container finished" podID="6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc" containerID="1215852094aff592dcbf374769e6d71692e8f6e30c02cab5c55382c5d14210a1" exitCode=0 Oct 08 00:22:11 crc kubenswrapper[4967]: I1008 00:22:11.358161 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-2-build" event={"ID":"6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc","Type":"ContainerDied","Data":"1215852094aff592dcbf374769e6d71692e8f6e30c02cab5c55382c5d14210a1"} Oct 08 00:22:12 crc kubenswrapper[4967]: I1008 00:22:12.365107 4967 generic.go:334] "Generic (PLEG): container finished" podID="6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc" containerID="9c45fa6cc4ec4a900aebe309274ed290bce8fe2fcec4e8711d61d9faf37e894a" exitCode=0 Oct 08 00:22:12 crc kubenswrapper[4967]: I1008 00:22:12.365177 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-2-build" event={"ID":"6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc","Type":"ContainerDied","Data":"9c45fa6cc4ec4a900aebe309274ed290bce8fe2fcec4e8711d61d9faf37e894a"} Oct 08 00:22:12 crc kubenswrapper[4967]: I1008 00:22:12.402414 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_sg-core-2-build_6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc/manage-dockerfile/0.log" Oct 08 00:22:13 crc kubenswrapper[4967]: I1008 00:22:13.376478 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-2-build" event={"ID":"6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc","Type":"ContainerStarted","Data":"c39117182dfe2e4f1b0e3fee21dc1a022f9d1423b33bd06bdcbe9d116332204e"} Oct 08 00:22:13 crc kubenswrapper[4967]: I1008 00:22:13.408575 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/sg-core-2-build" podStartSLOduration=4.4085568760000005 podStartE2EDuration="4.408556876s" podCreationTimestamp="2025-10-08 00:22:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:22:13.404940745 +0000 UTC m=+932.055475414" watchObservedRunningTime="2025-10-08 00:22:13.408556876 +0000 UTC m=+932.059091525" Oct 08 00:23:06 crc kubenswrapper[4967]: I1008 00:23:06.920712 4967 patch_prober.go:28] interesting pod/machine-config-daemon-xvdtp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 00:23:06 crc kubenswrapper[4967]: I1008 00:23:06.921356 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" podUID="9461162a-c27d-465e-9ca7-2d3936877cd8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 00:23:36 crc kubenswrapper[4967]: I1008 00:23:36.920730 4967 patch_prober.go:28] interesting pod/machine-config-daemon-xvdtp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 00:23:36 crc kubenswrapper[4967]: I1008 00:23:36.921580 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" podUID="9461162a-c27d-465e-9ca7-2d3936877cd8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 00:24:06 crc kubenswrapper[4967]: I1008 00:24:06.920056 4967 patch_prober.go:28] interesting pod/machine-config-daemon-xvdtp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 00:24:06 crc kubenswrapper[4967]: I1008 00:24:06.922290 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" podUID="9461162a-c27d-465e-9ca7-2d3936877cd8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 00:24:06 crc kubenswrapper[4967]: I1008 00:24:06.922547 4967 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" Oct 08 00:24:06 crc kubenswrapper[4967]: I1008 00:24:06.923511 4967 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"bd8678b8254640836f0db760c5c6a0da933718c28fc5033465f9875f7bd725b1"} pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 00:24:06 crc kubenswrapper[4967]: I1008 00:24:06.923981 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" podUID="9461162a-c27d-465e-9ca7-2d3936877cd8" containerName="machine-config-daemon" containerID="cri-o://bd8678b8254640836f0db760c5c6a0da933718c28fc5033465f9875f7bd725b1" gracePeriod=600 Oct 08 00:24:08 crc kubenswrapper[4967]: I1008 00:24:08.262214 4967 generic.go:334] "Generic (PLEG): container finished" podID="9461162a-c27d-465e-9ca7-2d3936877cd8" containerID="bd8678b8254640836f0db760c5c6a0da933718c28fc5033465f9875f7bd725b1" exitCode=0 Oct 08 00:24:08 crc kubenswrapper[4967]: I1008 00:24:08.262299 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" event={"ID":"9461162a-c27d-465e-9ca7-2d3936877cd8","Type":"ContainerDied","Data":"bd8678b8254640836f0db760c5c6a0da933718c28fc5033465f9875f7bd725b1"} Oct 08 00:24:08 crc kubenswrapper[4967]: I1008 00:24:08.262552 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" event={"ID":"9461162a-c27d-465e-9ca7-2d3936877cd8","Type":"ContainerStarted","Data":"a9bcd32e3405379dbe88886537687c4d0982cbf18aecd6bc1e1ae62026f21458"} Oct 08 00:24:08 crc kubenswrapper[4967]: I1008 00:24:08.262574 4967 scope.go:117] "RemoveContainer" containerID="312ed5e75849b65841b141a338e52c38e6c54f84a2e110368c62a60516be0c6b" Oct 08 00:25:32 crc kubenswrapper[4967]: I1008 00:25:32.848326 4967 generic.go:334] "Generic (PLEG): container finished" podID="6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc" containerID="c39117182dfe2e4f1b0e3fee21dc1a022f9d1423b33bd06bdcbe9d116332204e" exitCode=0 Oct 08 00:25:32 crc kubenswrapper[4967]: I1008 00:25:32.848436 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-2-build" event={"ID":"6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc","Type":"ContainerDied","Data":"c39117182dfe2e4f1b0e3fee21dc1a022f9d1423b33bd06bdcbe9d116332204e"} Oct 08 00:25:34 crc kubenswrapper[4967]: I1008 00:25:34.154483 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-core-2-build" Oct 08 00:25:34 crc kubenswrapper[4967]: I1008 00:25:34.246243 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc-build-blob-cache\") pod \"6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc\" (UID: \"6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc\") " Oct 08 00:25:34 crc kubenswrapper[4967]: I1008 00:25:34.246340 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-r7snr-push\" (UniqueName: \"kubernetes.io/secret/6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc-builder-dockercfg-r7snr-push\") pod \"6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc\" (UID: \"6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc\") " Oct 08 00:25:34 crc kubenswrapper[4967]: I1008 00:25:34.246380 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6b88q\" (UniqueName: \"kubernetes.io/projected/6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc-kube-api-access-6b88q\") pod \"6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc\" (UID: \"6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc\") " Oct 08 00:25:34 crc kubenswrapper[4967]: I1008 00:25:34.246404 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc-container-storage-root\") pod \"6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc\" (UID: \"6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc\") " Oct 08 00:25:34 crc kubenswrapper[4967]: I1008 00:25:34.246442 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc-buildworkdir\") pod \"6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc\" (UID: \"6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc\") " Oct 08 00:25:34 crc kubenswrapper[4967]: I1008 00:25:34.246499 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc-container-storage-run\") pod \"6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc\" (UID: \"6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc\") " Oct 08 00:25:34 crc kubenswrapper[4967]: I1008 00:25:34.246539 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-r7snr-pull\" (UniqueName: \"kubernetes.io/secret/6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc-builder-dockercfg-r7snr-pull\") pod \"6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc\" (UID: \"6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc\") " Oct 08 00:25:34 crc kubenswrapper[4967]: I1008 00:25:34.246560 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc-node-pullsecrets\") pod \"6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc\" (UID: \"6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc\") " Oct 08 00:25:34 crc kubenswrapper[4967]: I1008 00:25:34.246582 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc-build-proxy-ca-bundles\") pod \"6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc\" (UID: \"6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc\") " Oct 08 00:25:34 crc kubenswrapper[4967]: I1008 00:25:34.246612 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc-build-ca-bundles\") pod \"6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc\" (UID: \"6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc\") " Oct 08 00:25:34 crc kubenswrapper[4967]: I1008 00:25:34.246648 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc-build-system-configs\") pod \"6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc\" (UID: \"6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc\") " Oct 08 00:25:34 crc kubenswrapper[4967]: I1008 00:25:34.246688 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc-buildcachedir\") pod \"6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc\" (UID: \"6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc\") " Oct 08 00:25:34 crc kubenswrapper[4967]: I1008 00:25:34.247068 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc" (UID: "6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 00:25:34 crc kubenswrapper[4967]: I1008 00:25:34.247061 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc" (UID: "6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 00:25:34 crc kubenswrapper[4967]: I1008 00:25:34.247798 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc" (UID: "6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:25:34 crc kubenswrapper[4967]: I1008 00:25:34.248997 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc" (UID: "6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:25:34 crc kubenswrapper[4967]: I1008 00:25:34.249706 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc" (UID: "6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:25:34 crc kubenswrapper[4967]: I1008 00:25:34.251340 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc" (UID: "6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:25:34 crc kubenswrapper[4967]: I1008 00:25:34.251505 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc-builder-dockercfg-r7snr-push" (OuterVolumeSpecName: "builder-dockercfg-r7snr-push") pod "6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc" (UID: "6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc"). InnerVolumeSpecName "builder-dockercfg-r7snr-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:25:34 crc kubenswrapper[4967]: I1008 00:25:34.252650 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc-kube-api-access-6b88q" (OuterVolumeSpecName: "kube-api-access-6b88q") pod "6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc" (UID: "6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc"). InnerVolumeSpecName "kube-api-access-6b88q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:25:34 crc kubenswrapper[4967]: I1008 00:25:34.254516 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc-builder-dockercfg-r7snr-pull" (OuterVolumeSpecName: "builder-dockercfg-r7snr-pull") pod "6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc" (UID: "6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc"). InnerVolumeSpecName "builder-dockercfg-r7snr-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:25:34 crc kubenswrapper[4967]: I1008 00:25:34.262350 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc" (UID: "6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:25:34 crc kubenswrapper[4967]: I1008 00:25:34.348572 4967 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-r7snr-pull\" (UniqueName: \"kubernetes.io/secret/6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc-builder-dockercfg-r7snr-pull\") on node \"crc\" DevicePath \"\"" Oct 08 00:25:34 crc kubenswrapper[4967]: I1008 00:25:34.348645 4967 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Oct 08 00:25:34 crc kubenswrapper[4967]: I1008 00:25:34.348708 4967 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 08 00:25:34 crc kubenswrapper[4967]: I1008 00:25:34.348735 4967 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 08 00:25:34 crc kubenswrapper[4967]: I1008 00:25:34.348759 4967 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc-build-system-configs\") on node \"crc\" DevicePath \"\"" Oct 08 00:25:34 crc kubenswrapper[4967]: I1008 00:25:34.348782 4967 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc-buildcachedir\") on node \"crc\" DevicePath \"\"" Oct 08 00:25:34 crc kubenswrapper[4967]: I1008 00:25:34.348801 4967 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-r7snr-push\" (UniqueName: \"kubernetes.io/secret/6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc-builder-dockercfg-r7snr-push\") on node \"crc\" DevicePath \"\"" Oct 08 00:25:34 crc kubenswrapper[4967]: I1008 00:25:34.348819 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6b88q\" (UniqueName: \"kubernetes.io/projected/6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc-kube-api-access-6b88q\") on node \"crc\" DevicePath \"\"" Oct 08 00:25:34 crc kubenswrapper[4967]: I1008 00:25:34.348837 4967 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc-buildworkdir\") on node \"crc\" DevicePath \"\"" Oct 08 00:25:34 crc kubenswrapper[4967]: I1008 00:25:34.348857 4967 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc-container-storage-run\") on node \"crc\" DevicePath \"\"" Oct 08 00:25:34 crc kubenswrapper[4967]: I1008 00:25:34.641874 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc" (UID: "6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:25:34 crc kubenswrapper[4967]: I1008 00:25:34.653832 4967 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc-build-blob-cache\") on node \"crc\" DevicePath \"\"" Oct 08 00:25:34 crc kubenswrapper[4967]: I1008 00:25:34.867532 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-core-2-build" event={"ID":"6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc","Type":"ContainerDied","Data":"6334261af191189aa20688c2e6e805f2551dcab6688c2d544163a650ff8ae299"} Oct 08 00:25:34 crc kubenswrapper[4967]: I1008 00:25:34.867579 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6334261af191189aa20688c2e6e805f2551dcab6688c2d544163a650ff8ae299" Oct 08 00:25:34 crc kubenswrapper[4967]: I1008 00:25:34.867641 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-core-2-build" Oct 08 00:25:37 crc kubenswrapper[4967]: I1008 00:25:37.369492 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc" (UID: "6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:25:37 crc kubenswrapper[4967]: I1008 00:25:37.393895 4967 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc-container-storage-root\") on node \"crc\" DevicePath \"\"" Oct 08 00:25:38 crc kubenswrapper[4967]: I1008 00:25:38.988469 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/sg-bridge-1-build"] Oct 08 00:25:38 crc kubenswrapper[4967]: E1008 00:25:38.989011 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc" containerName="git-clone" Oct 08 00:25:38 crc kubenswrapper[4967]: I1008 00:25:38.989036 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc" containerName="git-clone" Oct 08 00:25:38 crc kubenswrapper[4967]: E1008 00:25:38.989050 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc" containerName="manage-dockerfile" Oct 08 00:25:38 crc kubenswrapper[4967]: I1008 00:25:38.989058 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc" containerName="manage-dockerfile" Oct 08 00:25:38 crc kubenswrapper[4967]: E1008 00:25:38.989069 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc" containerName="docker-build" Oct 08 00:25:38 crc kubenswrapper[4967]: I1008 00:25:38.989079 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc" containerName="docker-build" Oct 08 00:25:38 crc kubenswrapper[4967]: I1008 00:25:38.989210 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d02c04e-3a4d-405c-8dca-c9f7aaeb82bc" containerName="docker-build" Oct 08 00:25:38 crc kubenswrapper[4967]: I1008 00:25:38.989937 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-bridge-1-build" Oct 08 00:25:38 crc kubenswrapper[4967]: I1008 00:25:38.992800 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"sg-bridge-1-ca" Oct 08 00:25:38 crc kubenswrapper[4967]: I1008 00:25:38.992801 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"sg-bridge-1-sys-config" Oct 08 00:25:38 crc kubenswrapper[4967]: I1008 00:25:38.992967 4967 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-r7snr" Oct 08 00:25:38 crc kubenswrapper[4967]: I1008 00:25:38.996197 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"sg-bridge-1-global-ca" Oct 08 00:25:39 crc kubenswrapper[4967]: I1008 00:25:39.014875 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/sg-bridge-1-build"] Oct 08 00:25:39 crc kubenswrapper[4967]: I1008 00:25:39.015047 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/6cc52f44-a91c-45d1-b86a-7e658bf4cd5b-container-storage-root\") pod \"sg-bridge-1-build\" (UID: \"6cc52f44-a91c-45d1-b86a-7e658bf4cd5b\") " pod="service-telemetry/sg-bridge-1-build" Oct 08 00:25:39 crc kubenswrapper[4967]: I1008 00:25:39.015150 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/6cc52f44-a91c-45d1-b86a-7e658bf4cd5b-build-blob-cache\") pod \"sg-bridge-1-build\" (UID: \"6cc52f44-a91c-45d1-b86a-7e658bf4cd5b\") " pod="service-telemetry/sg-bridge-1-build" Oct 08 00:25:39 crc kubenswrapper[4967]: I1008 00:25:39.015249 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/6cc52f44-a91c-45d1-b86a-7e658bf4cd5b-buildcachedir\") pod \"sg-bridge-1-build\" (UID: \"6cc52f44-a91c-45d1-b86a-7e658bf4cd5b\") " pod="service-telemetry/sg-bridge-1-build" Oct 08 00:25:39 crc kubenswrapper[4967]: I1008 00:25:39.015288 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dznw7\" (UniqueName: \"kubernetes.io/projected/6cc52f44-a91c-45d1-b86a-7e658bf4cd5b-kube-api-access-dznw7\") pod \"sg-bridge-1-build\" (UID: \"6cc52f44-a91c-45d1-b86a-7e658bf4cd5b\") " pod="service-telemetry/sg-bridge-1-build" Oct 08 00:25:39 crc kubenswrapper[4967]: I1008 00:25:39.015375 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/6cc52f44-a91c-45d1-b86a-7e658bf4cd5b-container-storage-run\") pod \"sg-bridge-1-build\" (UID: \"6cc52f44-a91c-45d1-b86a-7e658bf4cd5b\") " pod="service-telemetry/sg-bridge-1-build" Oct 08 00:25:39 crc kubenswrapper[4967]: I1008 00:25:39.015490 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/6cc52f44-a91c-45d1-b86a-7e658bf4cd5b-node-pullsecrets\") pod \"sg-bridge-1-build\" (UID: \"6cc52f44-a91c-45d1-b86a-7e658bf4cd5b\") " pod="service-telemetry/sg-bridge-1-build" Oct 08 00:25:39 crc kubenswrapper[4967]: I1008 00:25:39.015562 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/6cc52f44-a91c-45d1-b86a-7e658bf4cd5b-buildworkdir\") pod \"sg-bridge-1-build\" (UID: \"6cc52f44-a91c-45d1-b86a-7e658bf4cd5b\") " pod="service-telemetry/sg-bridge-1-build" Oct 08 00:25:39 crc kubenswrapper[4967]: I1008 00:25:39.015597 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-r7snr-push\" (UniqueName: \"kubernetes.io/secret/6cc52f44-a91c-45d1-b86a-7e658bf4cd5b-builder-dockercfg-r7snr-push\") pod \"sg-bridge-1-build\" (UID: \"6cc52f44-a91c-45d1-b86a-7e658bf4cd5b\") " pod="service-telemetry/sg-bridge-1-build" Oct 08 00:25:39 crc kubenswrapper[4967]: I1008 00:25:39.015643 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6cc52f44-a91c-45d1-b86a-7e658bf4cd5b-build-proxy-ca-bundles\") pod \"sg-bridge-1-build\" (UID: \"6cc52f44-a91c-45d1-b86a-7e658bf4cd5b\") " pod="service-telemetry/sg-bridge-1-build" Oct 08 00:25:39 crc kubenswrapper[4967]: I1008 00:25:39.015788 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-r7snr-pull\" (UniqueName: \"kubernetes.io/secret/6cc52f44-a91c-45d1-b86a-7e658bf4cd5b-builder-dockercfg-r7snr-pull\") pod \"sg-bridge-1-build\" (UID: \"6cc52f44-a91c-45d1-b86a-7e658bf4cd5b\") " pod="service-telemetry/sg-bridge-1-build" Oct 08 00:25:39 crc kubenswrapper[4967]: I1008 00:25:39.015867 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6cc52f44-a91c-45d1-b86a-7e658bf4cd5b-build-ca-bundles\") pod \"sg-bridge-1-build\" (UID: \"6cc52f44-a91c-45d1-b86a-7e658bf4cd5b\") " pod="service-telemetry/sg-bridge-1-build" Oct 08 00:25:39 crc kubenswrapper[4967]: I1008 00:25:39.015937 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/6cc52f44-a91c-45d1-b86a-7e658bf4cd5b-build-system-configs\") pod \"sg-bridge-1-build\" (UID: \"6cc52f44-a91c-45d1-b86a-7e658bf4cd5b\") " pod="service-telemetry/sg-bridge-1-build" Oct 08 00:25:39 crc kubenswrapper[4967]: I1008 00:25:39.116892 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/6cc52f44-a91c-45d1-b86a-7e658bf4cd5b-container-storage-run\") pod \"sg-bridge-1-build\" (UID: \"6cc52f44-a91c-45d1-b86a-7e658bf4cd5b\") " pod="service-telemetry/sg-bridge-1-build" Oct 08 00:25:39 crc kubenswrapper[4967]: I1008 00:25:39.116940 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/6cc52f44-a91c-45d1-b86a-7e658bf4cd5b-node-pullsecrets\") pod \"sg-bridge-1-build\" (UID: \"6cc52f44-a91c-45d1-b86a-7e658bf4cd5b\") " pod="service-telemetry/sg-bridge-1-build" Oct 08 00:25:39 crc kubenswrapper[4967]: I1008 00:25:39.116966 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/6cc52f44-a91c-45d1-b86a-7e658bf4cd5b-buildworkdir\") pod \"sg-bridge-1-build\" (UID: \"6cc52f44-a91c-45d1-b86a-7e658bf4cd5b\") " pod="service-telemetry/sg-bridge-1-build" Oct 08 00:25:39 crc kubenswrapper[4967]: I1008 00:25:39.116979 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-r7snr-push\" (UniqueName: \"kubernetes.io/secret/6cc52f44-a91c-45d1-b86a-7e658bf4cd5b-builder-dockercfg-r7snr-push\") pod \"sg-bridge-1-build\" (UID: \"6cc52f44-a91c-45d1-b86a-7e658bf4cd5b\") " pod="service-telemetry/sg-bridge-1-build" Oct 08 00:25:39 crc kubenswrapper[4967]: I1008 00:25:39.117001 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6cc52f44-a91c-45d1-b86a-7e658bf4cd5b-build-proxy-ca-bundles\") pod \"sg-bridge-1-build\" (UID: \"6cc52f44-a91c-45d1-b86a-7e658bf4cd5b\") " pod="service-telemetry/sg-bridge-1-build" Oct 08 00:25:39 crc kubenswrapper[4967]: I1008 00:25:39.117021 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-r7snr-pull\" (UniqueName: \"kubernetes.io/secret/6cc52f44-a91c-45d1-b86a-7e658bf4cd5b-builder-dockercfg-r7snr-pull\") pod \"sg-bridge-1-build\" (UID: \"6cc52f44-a91c-45d1-b86a-7e658bf4cd5b\") " pod="service-telemetry/sg-bridge-1-build" Oct 08 00:25:39 crc kubenswrapper[4967]: I1008 00:25:39.117042 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6cc52f44-a91c-45d1-b86a-7e658bf4cd5b-build-ca-bundles\") pod \"sg-bridge-1-build\" (UID: \"6cc52f44-a91c-45d1-b86a-7e658bf4cd5b\") " pod="service-telemetry/sg-bridge-1-build" Oct 08 00:25:39 crc kubenswrapper[4967]: I1008 00:25:39.117061 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/6cc52f44-a91c-45d1-b86a-7e658bf4cd5b-build-system-configs\") pod \"sg-bridge-1-build\" (UID: \"6cc52f44-a91c-45d1-b86a-7e658bf4cd5b\") " pod="service-telemetry/sg-bridge-1-build" Oct 08 00:25:39 crc kubenswrapper[4967]: I1008 00:25:39.117087 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/6cc52f44-a91c-45d1-b86a-7e658bf4cd5b-container-storage-root\") pod \"sg-bridge-1-build\" (UID: \"6cc52f44-a91c-45d1-b86a-7e658bf4cd5b\") " pod="service-telemetry/sg-bridge-1-build" Oct 08 00:25:39 crc kubenswrapper[4967]: I1008 00:25:39.117110 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/6cc52f44-a91c-45d1-b86a-7e658bf4cd5b-build-blob-cache\") pod \"sg-bridge-1-build\" (UID: \"6cc52f44-a91c-45d1-b86a-7e658bf4cd5b\") " pod="service-telemetry/sg-bridge-1-build" Oct 08 00:25:39 crc kubenswrapper[4967]: I1008 00:25:39.117126 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/6cc52f44-a91c-45d1-b86a-7e658bf4cd5b-buildcachedir\") pod \"sg-bridge-1-build\" (UID: \"6cc52f44-a91c-45d1-b86a-7e658bf4cd5b\") " pod="service-telemetry/sg-bridge-1-build" Oct 08 00:25:39 crc kubenswrapper[4967]: I1008 00:25:39.117142 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dznw7\" (UniqueName: \"kubernetes.io/projected/6cc52f44-a91c-45d1-b86a-7e658bf4cd5b-kube-api-access-dznw7\") pod \"sg-bridge-1-build\" (UID: \"6cc52f44-a91c-45d1-b86a-7e658bf4cd5b\") " pod="service-telemetry/sg-bridge-1-build" Oct 08 00:25:39 crc kubenswrapper[4967]: I1008 00:25:39.117239 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/6cc52f44-a91c-45d1-b86a-7e658bf4cd5b-node-pullsecrets\") pod \"sg-bridge-1-build\" (UID: \"6cc52f44-a91c-45d1-b86a-7e658bf4cd5b\") " pod="service-telemetry/sg-bridge-1-build" Oct 08 00:25:39 crc kubenswrapper[4967]: I1008 00:25:39.117379 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/6cc52f44-a91c-45d1-b86a-7e658bf4cd5b-buildcachedir\") pod \"sg-bridge-1-build\" (UID: \"6cc52f44-a91c-45d1-b86a-7e658bf4cd5b\") " pod="service-telemetry/sg-bridge-1-build" Oct 08 00:25:39 crc kubenswrapper[4967]: I1008 00:25:39.117432 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/6cc52f44-a91c-45d1-b86a-7e658bf4cd5b-container-storage-run\") pod \"sg-bridge-1-build\" (UID: \"6cc52f44-a91c-45d1-b86a-7e658bf4cd5b\") " pod="service-telemetry/sg-bridge-1-build" Oct 08 00:25:39 crc kubenswrapper[4967]: I1008 00:25:39.117540 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/6cc52f44-a91c-45d1-b86a-7e658bf4cd5b-buildworkdir\") pod \"sg-bridge-1-build\" (UID: \"6cc52f44-a91c-45d1-b86a-7e658bf4cd5b\") " pod="service-telemetry/sg-bridge-1-build" Oct 08 00:25:39 crc kubenswrapper[4967]: I1008 00:25:39.117738 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/6cc52f44-a91c-45d1-b86a-7e658bf4cd5b-container-storage-root\") pod \"sg-bridge-1-build\" (UID: \"6cc52f44-a91c-45d1-b86a-7e658bf4cd5b\") " pod="service-telemetry/sg-bridge-1-build" Oct 08 00:25:39 crc kubenswrapper[4967]: I1008 00:25:39.117816 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/6cc52f44-a91c-45d1-b86a-7e658bf4cd5b-build-blob-cache\") pod \"sg-bridge-1-build\" (UID: \"6cc52f44-a91c-45d1-b86a-7e658bf4cd5b\") " pod="service-telemetry/sg-bridge-1-build" Oct 08 00:25:39 crc kubenswrapper[4967]: I1008 00:25:39.118465 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6cc52f44-a91c-45d1-b86a-7e658bf4cd5b-build-proxy-ca-bundles\") pod \"sg-bridge-1-build\" (UID: \"6cc52f44-a91c-45d1-b86a-7e658bf4cd5b\") " pod="service-telemetry/sg-bridge-1-build" Oct 08 00:25:39 crc kubenswrapper[4967]: I1008 00:25:39.118692 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/6cc52f44-a91c-45d1-b86a-7e658bf4cd5b-build-system-configs\") pod \"sg-bridge-1-build\" (UID: \"6cc52f44-a91c-45d1-b86a-7e658bf4cd5b\") " pod="service-telemetry/sg-bridge-1-build" Oct 08 00:25:39 crc kubenswrapper[4967]: I1008 00:25:39.118789 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6cc52f44-a91c-45d1-b86a-7e658bf4cd5b-build-ca-bundles\") pod \"sg-bridge-1-build\" (UID: \"6cc52f44-a91c-45d1-b86a-7e658bf4cd5b\") " pod="service-telemetry/sg-bridge-1-build" Oct 08 00:25:39 crc kubenswrapper[4967]: I1008 00:25:39.123506 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-r7snr-pull\" (UniqueName: \"kubernetes.io/secret/6cc52f44-a91c-45d1-b86a-7e658bf4cd5b-builder-dockercfg-r7snr-pull\") pod \"sg-bridge-1-build\" (UID: \"6cc52f44-a91c-45d1-b86a-7e658bf4cd5b\") " pod="service-telemetry/sg-bridge-1-build" Oct 08 00:25:39 crc kubenswrapper[4967]: I1008 00:25:39.128241 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-r7snr-push\" (UniqueName: \"kubernetes.io/secret/6cc52f44-a91c-45d1-b86a-7e658bf4cd5b-builder-dockercfg-r7snr-push\") pod \"sg-bridge-1-build\" (UID: \"6cc52f44-a91c-45d1-b86a-7e658bf4cd5b\") " pod="service-telemetry/sg-bridge-1-build" Oct 08 00:25:39 crc kubenswrapper[4967]: I1008 00:25:39.137898 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dznw7\" (UniqueName: \"kubernetes.io/projected/6cc52f44-a91c-45d1-b86a-7e658bf4cd5b-kube-api-access-dznw7\") pod \"sg-bridge-1-build\" (UID: \"6cc52f44-a91c-45d1-b86a-7e658bf4cd5b\") " pod="service-telemetry/sg-bridge-1-build" Oct 08 00:25:39 crc kubenswrapper[4967]: I1008 00:25:39.313055 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-bridge-1-build" Oct 08 00:25:39 crc kubenswrapper[4967]: I1008 00:25:39.615430 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/sg-bridge-1-build"] Oct 08 00:25:39 crc kubenswrapper[4967]: I1008 00:25:39.905650 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-1-build" event={"ID":"6cc52f44-a91c-45d1-b86a-7e658bf4cd5b","Type":"ContainerStarted","Data":"02e3f8df4e964e4e5637cff64263bdb55dabf7c98eb18110ca86f5ca54bd4343"} Oct 08 00:25:40 crc kubenswrapper[4967]: I1008 00:25:40.915903 4967 generic.go:334] "Generic (PLEG): container finished" podID="6cc52f44-a91c-45d1-b86a-7e658bf4cd5b" containerID="b61d29d4dc3065525df66102951ebde9347d69c4aad384cd158f9575fa987ea8" exitCode=0 Oct 08 00:25:40 crc kubenswrapper[4967]: I1008 00:25:40.916005 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-1-build" event={"ID":"6cc52f44-a91c-45d1-b86a-7e658bf4cd5b","Type":"ContainerDied","Data":"b61d29d4dc3065525df66102951ebde9347d69c4aad384cd158f9575fa987ea8"} Oct 08 00:25:41 crc kubenswrapper[4967]: I1008 00:25:41.926785 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-1-build" event={"ID":"6cc52f44-a91c-45d1-b86a-7e658bf4cd5b","Type":"ContainerStarted","Data":"ddec5842671230989b43b33ed0074978a284dec734622e318616a56c738ed951"} Oct 08 00:25:41 crc kubenswrapper[4967]: I1008 00:25:41.972498 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/sg-bridge-1-build" podStartSLOduration=3.972469907 podStartE2EDuration="3.972469907s" podCreationTimestamp="2025-10-08 00:25:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:25:41.966161238 +0000 UTC m=+1140.616695957" watchObservedRunningTime="2025-10-08 00:25:41.972469907 +0000 UTC m=+1140.623004596" Oct 08 00:25:49 crc kubenswrapper[4967]: I1008 00:25:49.738708 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/sg-bridge-1-build"] Oct 08 00:25:49 crc kubenswrapper[4967]: I1008 00:25:49.739452 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="service-telemetry/sg-bridge-1-build" podUID="6cc52f44-a91c-45d1-b86a-7e658bf4cd5b" containerName="docker-build" containerID="cri-o://ddec5842671230989b43b33ed0074978a284dec734622e318616a56c738ed951" gracePeriod=30 Oct 08 00:25:49 crc kubenswrapper[4967]: I1008 00:25:49.987114 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_sg-bridge-1-build_6cc52f44-a91c-45d1-b86a-7e658bf4cd5b/docker-build/0.log" Oct 08 00:25:49 crc kubenswrapper[4967]: I1008 00:25:49.988147 4967 generic.go:334] "Generic (PLEG): container finished" podID="6cc52f44-a91c-45d1-b86a-7e658bf4cd5b" containerID="ddec5842671230989b43b33ed0074978a284dec734622e318616a56c738ed951" exitCode=1 Oct 08 00:25:49 crc kubenswrapper[4967]: I1008 00:25:49.988214 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-1-build" event={"ID":"6cc52f44-a91c-45d1-b86a-7e658bf4cd5b","Type":"ContainerDied","Data":"ddec5842671230989b43b33ed0074978a284dec734622e318616a56c738ed951"} Oct 08 00:25:49 crc kubenswrapper[4967]: I1008 00:25:49.988255 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-1-build" event={"ID":"6cc52f44-a91c-45d1-b86a-7e658bf4cd5b","Type":"ContainerDied","Data":"02e3f8df4e964e4e5637cff64263bdb55dabf7c98eb18110ca86f5ca54bd4343"} Oct 08 00:25:49 crc kubenswrapper[4967]: I1008 00:25:49.988281 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="02e3f8df4e964e4e5637cff64263bdb55dabf7c98eb18110ca86f5ca54bd4343" Oct 08 00:25:50 crc kubenswrapper[4967]: I1008 00:25:50.035034 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_sg-bridge-1-build_6cc52f44-a91c-45d1-b86a-7e658bf4cd5b/docker-build/0.log" Oct 08 00:25:50 crc kubenswrapper[4967]: I1008 00:25:50.035800 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-bridge-1-build" Oct 08 00:25:50 crc kubenswrapper[4967]: I1008 00:25:50.082014 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/6cc52f44-a91c-45d1-b86a-7e658bf4cd5b-container-storage-root\") pod \"6cc52f44-a91c-45d1-b86a-7e658bf4cd5b\" (UID: \"6cc52f44-a91c-45d1-b86a-7e658bf4cd5b\") " Oct 08 00:25:50 crc kubenswrapper[4967]: I1008 00:25:50.082119 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6cc52f44-a91c-45d1-b86a-7e658bf4cd5b-build-proxy-ca-bundles\") pod \"6cc52f44-a91c-45d1-b86a-7e658bf4cd5b\" (UID: \"6cc52f44-a91c-45d1-b86a-7e658bf4cd5b\") " Oct 08 00:25:50 crc kubenswrapper[4967]: I1008 00:25:50.082174 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/6cc52f44-a91c-45d1-b86a-7e658bf4cd5b-node-pullsecrets\") pod \"6cc52f44-a91c-45d1-b86a-7e658bf4cd5b\" (UID: \"6cc52f44-a91c-45d1-b86a-7e658bf4cd5b\") " Oct 08 00:25:50 crc kubenswrapper[4967]: I1008 00:25:50.082248 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-r7snr-push\" (UniqueName: \"kubernetes.io/secret/6cc52f44-a91c-45d1-b86a-7e658bf4cd5b-builder-dockercfg-r7snr-push\") pod \"6cc52f44-a91c-45d1-b86a-7e658bf4cd5b\" (UID: \"6cc52f44-a91c-45d1-b86a-7e658bf4cd5b\") " Oct 08 00:25:50 crc kubenswrapper[4967]: I1008 00:25:50.082308 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6cc52f44-a91c-45d1-b86a-7e658bf4cd5b-build-ca-bundles\") pod \"6cc52f44-a91c-45d1-b86a-7e658bf4cd5b\" (UID: \"6cc52f44-a91c-45d1-b86a-7e658bf4cd5b\") " Oct 08 00:25:50 crc kubenswrapper[4967]: I1008 00:25:50.082365 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/6cc52f44-a91c-45d1-b86a-7e658bf4cd5b-build-blob-cache\") pod \"6cc52f44-a91c-45d1-b86a-7e658bf4cd5b\" (UID: \"6cc52f44-a91c-45d1-b86a-7e658bf4cd5b\") " Oct 08 00:25:50 crc kubenswrapper[4967]: I1008 00:25:50.082405 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/6cc52f44-a91c-45d1-b86a-7e658bf4cd5b-buildworkdir\") pod \"6cc52f44-a91c-45d1-b86a-7e658bf4cd5b\" (UID: \"6cc52f44-a91c-45d1-b86a-7e658bf4cd5b\") " Oct 08 00:25:50 crc kubenswrapper[4967]: I1008 00:25:50.082474 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dznw7\" (UniqueName: \"kubernetes.io/projected/6cc52f44-a91c-45d1-b86a-7e658bf4cd5b-kube-api-access-dznw7\") pod \"6cc52f44-a91c-45d1-b86a-7e658bf4cd5b\" (UID: \"6cc52f44-a91c-45d1-b86a-7e658bf4cd5b\") " Oct 08 00:25:50 crc kubenswrapper[4967]: I1008 00:25:50.082563 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-r7snr-pull\" (UniqueName: \"kubernetes.io/secret/6cc52f44-a91c-45d1-b86a-7e658bf4cd5b-builder-dockercfg-r7snr-pull\") pod \"6cc52f44-a91c-45d1-b86a-7e658bf4cd5b\" (UID: \"6cc52f44-a91c-45d1-b86a-7e658bf4cd5b\") " Oct 08 00:25:50 crc kubenswrapper[4967]: I1008 00:25:50.082638 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/6cc52f44-a91c-45d1-b86a-7e658bf4cd5b-buildcachedir\") pod \"6cc52f44-a91c-45d1-b86a-7e658bf4cd5b\" (UID: \"6cc52f44-a91c-45d1-b86a-7e658bf4cd5b\") " Oct 08 00:25:50 crc kubenswrapper[4967]: I1008 00:25:50.082711 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/6cc52f44-a91c-45d1-b86a-7e658bf4cd5b-build-system-configs\") pod \"6cc52f44-a91c-45d1-b86a-7e658bf4cd5b\" (UID: \"6cc52f44-a91c-45d1-b86a-7e658bf4cd5b\") " Oct 08 00:25:50 crc kubenswrapper[4967]: I1008 00:25:50.082797 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/6cc52f44-a91c-45d1-b86a-7e658bf4cd5b-container-storage-run\") pod \"6cc52f44-a91c-45d1-b86a-7e658bf4cd5b\" (UID: \"6cc52f44-a91c-45d1-b86a-7e658bf4cd5b\") " Oct 08 00:25:50 crc kubenswrapper[4967]: I1008 00:25:50.084277 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6cc52f44-a91c-45d1-b86a-7e658bf4cd5b-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "6cc52f44-a91c-45d1-b86a-7e658bf4cd5b" (UID: "6cc52f44-a91c-45d1-b86a-7e658bf4cd5b"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 00:25:50 crc kubenswrapper[4967]: I1008 00:25:50.084503 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6cc52f44-a91c-45d1-b86a-7e658bf4cd5b-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "6cc52f44-a91c-45d1-b86a-7e658bf4cd5b" (UID: "6cc52f44-a91c-45d1-b86a-7e658bf4cd5b"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:25:50 crc kubenswrapper[4967]: I1008 00:25:50.084560 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6cc52f44-a91c-45d1-b86a-7e658bf4cd5b-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "6cc52f44-a91c-45d1-b86a-7e658bf4cd5b" (UID: "6cc52f44-a91c-45d1-b86a-7e658bf4cd5b"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 00:25:50 crc kubenswrapper[4967]: I1008 00:25:50.084924 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6cc52f44-a91c-45d1-b86a-7e658bf4cd5b-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "6cc52f44-a91c-45d1-b86a-7e658bf4cd5b" (UID: "6cc52f44-a91c-45d1-b86a-7e658bf4cd5b"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:25:50 crc kubenswrapper[4967]: I1008 00:25:50.085017 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6cc52f44-a91c-45d1-b86a-7e658bf4cd5b-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "6cc52f44-a91c-45d1-b86a-7e658bf4cd5b" (UID: "6cc52f44-a91c-45d1-b86a-7e658bf4cd5b"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:25:50 crc kubenswrapper[4967]: I1008 00:25:50.086010 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6cc52f44-a91c-45d1-b86a-7e658bf4cd5b-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "6cc52f44-a91c-45d1-b86a-7e658bf4cd5b" (UID: "6cc52f44-a91c-45d1-b86a-7e658bf4cd5b"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:25:50 crc kubenswrapper[4967]: I1008 00:25:50.088024 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6cc52f44-a91c-45d1-b86a-7e658bf4cd5b-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "6cc52f44-a91c-45d1-b86a-7e658bf4cd5b" (UID: "6cc52f44-a91c-45d1-b86a-7e658bf4cd5b"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:25:50 crc kubenswrapper[4967]: I1008 00:25:50.091898 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6cc52f44-a91c-45d1-b86a-7e658bf4cd5b-builder-dockercfg-r7snr-pull" (OuterVolumeSpecName: "builder-dockercfg-r7snr-pull") pod "6cc52f44-a91c-45d1-b86a-7e658bf4cd5b" (UID: "6cc52f44-a91c-45d1-b86a-7e658bf4cd5b"). InnerVolumeSpecName "builder-dockercfg-r7snr-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:25:50 crc kubenswrapper[4967]: I1008 00:25:50.092172 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6cc52f44-a91c-45d1-b86a-7e658bf4cd5b-kube-api-access-dznw7" (OuterVolumeSpecName: "kube-api-access-dznw7") pod "6cc52f44-a91c-45d1-b86a-7e658bf4cd5b" (UID: "6cc52f44-a91c-45d1-b86a-7e658bf4cd5b"). InnerVolumeSpecName "kube-api-access-dznw7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:25:50 crc kubenswrapper[4967]: I1008 00:25:50.093180 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6cc52f44-a91c-45d1-b86a-7e658bf4cd5b-builder-dockercfg-r7snr-push" (OuterVolumeSpecName: "builder-dockercfg-r7snr-push") pod "6cc52f44-a91c-45d1-b86a-7e658bf4cd5b" (UID: "6cc52f44-a91c-45d1-b86a-7e658bf4cd5b"). InnerVolumeSpecName "builder-dockercfg-r7snr-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:25:50 crc kubenswrapper[4967]: I1008 00:25:50.155288 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6cc52f44-a91c-45d1-b86a-7e658bf4cd5b-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "6cc52f44-a91c-45d1-b86a-7e658bf4cd5b" (UID: "6cc52f44-a91c-45d1-b86a-7e658bf4cd5b"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:25:50 crc kubenswrapper[4967]: I1008 00:25:50.184492 4967 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-r7snr-pull\" (UniqueName: \"kubernetes.io/secret/6cc52f44-a91c-45d1-b86a-7e658bf4cd5b-builder-dockercfg-r7snr-pull\") on node \"crc\" DevicePath \"\"" Oct 08 00:25:50 crc kubenswrapper[4967]: I1008 00:25:50.184531 4967 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/6cc52f44-a91c-45d1-b86a-7e658bf4cd5b-buildcachedir\") on node \"crc\" DevicePath \"\"" Oct 08 00:25:50 crc kubenswrapper[4967]: I1008 00:25:50.184544 4967 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/6cc52f44-a91c-45d1-b86a-7e658bf4cd5b-build-system-configs\") on node \"crc\" DevicePath \"\"" Oct 08 00:25:50 crc kubenswrapper[4967]: I1008 00:25:50.184557 4967 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/6cc52f44-a91c-45d1-b86a-7e658bf4cd5b-container-storage-run\") on node \"crc\" DevicePath \"\"" Oct 08 00:25:50 crc kubenswrapper[4967]: I1008 00:25:50.184579 4967 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6cc52f44-a91c-45d1-b86a-7e658bf4cd5b-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 08 00:25:50 crc kubenswrapper[4967]: I1008 00:25:50.184590 4967 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/6cc52f44-a91c-45d1-b86a-7e658bf4cd5b-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Oct 08 00:25:50 crc kubenswrapper[4967]: I1008 00:25:50.184599 4967 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-r7snr-push\" (UniqueName: \"kubernetes.io/secret/6cc52f44-a91c-45d1-b86a-7e658bf4cd5b-builder-dockercfg-r7snr-push\") on node \"crc\" DevicePath \"\"" Oct 08 00:25:50 crc kubenswrapper[4967]: I1008 00:25:50.184703 4967 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6cc52f44-a91c-45d1-b86a-7e658bf4cd5b-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 08 00:25:50 crc kubenswrapper[4967]: I1008 00:25:50.185524 4967 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/6cc52f44-a91c-45d1-b86a-7e658bf4cd5b-build-blob-cache\") on node \"crc\" DevicePath \"\"" Oct 08 00:25:50 crc kubenswrapper[4967]: I1008 00:25:50.185722 4967 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/6cc52f44-a91c-45d1-b86a-7e658bf4cd5b-buildworkdir\") on node \"crc\" DevicePath \"\"" Oct 08 00:25:50 crc kubenswrapper[4967]: I1008 00:25:50.185776 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dznw7\" (UniqueName: \"kubernetes.io/projected/6cc52f44-a91c-45d1-b86a-7e658bf4cd5b-kube-api-access-dznw7\") on node \"crc\" DevicePath \"\"" Oct 08 00:25:50 crc kubenswrapper[4967]: I1008 00:25:50.491489 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6cc52f44-a91c-45d1-b86a-7e658bf4cd5b-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "6cc52f44-a91c-45d1-b86a-7e658bf4cd5b" (UID: "6cc52f44-a91c-45d1-b86a-7e658bf4cd5b"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:25:50 crc kubenswrapper[4967]: I1008 00:25:50.592256 4967 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/6cc52f44-a91c-45d1-b86a-7e658bf4cd5b-container-storage-root\") on node \"crc\" DevicePath \"\"" Oct 08 00:25:50 crc kubenswrapper[4967]: I1008 00:25:50.995584 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-bridge-1-build" Oct 08 00:25:51 crc kubenswrapper[4967]: I1008 00:25:51.065748 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/sg-bridge-1-build"] Oct 08 00:25:51 crc kubenswrapper[4967]: I1008 00:25:51.071601 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["service-telemetry/sg-bridge-1-build"] Oct 08 00:25:51 crc kubenswrapper[4967]: I1008 00:25:51.348039 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/sg-bridge-2-build"] Oct 08 00:25:51 crc kubenswrapper[4967]: E1008 00:25:51.351041 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6cc52f44-a91c-45d1-b86a-7e658bf4cd5b" containerName="docker-build" Oct 08 00:25:51 crc kubenswrapper[4967]: I1008 00:25:51.351089 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="6cc52f44-a91c-45d1-b86a-7e658bf4cd5b" containerName="docker-build" Oct 08 00:25:51 crc kubenswrapper[4967]: E1008 00:25:51.351166 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6cc52f44-a91c-45d1-b86a-7e658bf4cd5b" containerName="manage-dockerfile" Oct 08 00:25:51 crc kubenswrapper[4967]: I1008 00:25:51.351184 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="6cc52f44-a91c-45d1-b86a-7e658bf4cd5b" containerName="manage-dockerfile" Oct 08 00:25:51 crc kubenswrapper[4967]: I1008 00:25:51.351418 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="6cc52f44-a91c-45d1-b86a-7e658bf4cd5b" containerName="docker-build" Oct 08 00:25:51 crc kubenswrapper[4967]: I1008 00:25:51.352977 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-bridge-2-build" Oct 08 00:25:51 crc kubenswrapper[4967]: I1008 00:25:51.355913 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"sg-bridge-2-ca" Oct 08 00:25:51 crc kubenswrapper[4967]: I1008 00:25:51.356519 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"sg-bridge-2-global-ca" Oct 08 00:25:51 crc kubenswrapper[4967]: I1008 00:25:51.357782 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"sg-bridge-2-sys-config" Oct 08 00:25:51 crc kubenswrapper[4967]: I1008 00:25:51.358016 4967 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-r7snr" Oct 08 00:25:51 crc kubenswrapper[4967]: I1008 00:25:51.361439 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/sg-bridge-2-build"] Oct 08 00:25:51 crc kubenswrapper[4967]: I1008 00:25:51.403500 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-r7snr-pull\" (UniqueName: \"kubernetes.io/secret/a016ce4f-2506-4886-b6d2-3a51ce64dab0-builder-dockercfg-r7snr-pull\") pod \"sg-bridge-2-build\" (UID: \"a016ce4f-2506-4886-b6d2-3a51ce64dab0\") " pod="service-telemetry/sg-bridge-2-build" Oct 08 00:25:51 crc kubenswrapper[4967]: I1008 00:25:51.403572 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a016ce4f-2506-4886-b6d2-3a51ce64dab0-build-proxy-ca-bundles\") pod \"sg-bridge-2-build\" (UID: \"a016ce4f-2506-4886-b6d2-3a51ce64dab0\") " pod="service-telemetry/sg-bridge-2-build" Oct 08 00:25:51 crc kubenswrapper[4967]: I1008 00:25:51.403652 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/a016ce4f-2506-4886-b6d2-3a51ce64dab0-build-blob-cache\") pod \"sg-bridge-2-build\" (UID: \"a016ce4f-2506-4886-b6d2-3a51ce64dab0\") " pod="service-telemetry/sg-bridge-2-build" Oct 08 00:25:51 crc kubenswrapper[4967]: I1008 00:25:51.403739 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/a016ce4f-2506-4886-b6d2-3a51ce64dab0-build-system-configs\") pod \"sg-bridge-2-build\" (UID: \"a016ce4f-2506-4886-b6d2-3a51ce64dab0\") " pod="service-telemetry/sg-bridge-2-build" Oct 08 00:25:51 crc kubenswrapper[4967]: I1008 00:25:51.403770 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/a016ce4f-2506-4886-b6d2-3a51ce64dab0-buildworkdir\") pod \"sg-bridge-2-build\" (UID: \"a016ce4f-2506-4886-b6d2-3a51ce64dab0\") " pod="service-telemetry/sg-bridge-2-build" Oct 08 00:25:51 crc kubenswrapper[4967]: I1008 00:25:51.403810 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/a016ce4f-2506-4886-b6d2-3a51ce64dab0-node-pullsecrets\") pod \"sg-bridge-2-build\" (UID: \"a016ce4f-2506-4886-b6d2-3a51ce64dab0\") " pod="service-telemetry/sg-bridge-2-build" Oct 08 00:25:51 crc kubenswrapper[4967]: I1008 00:25:51.403860 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x7752\" (UniqueName: \"kubernetes.io/projected/a016ce4f-2506-4886-b6d2-3a51ce64dab0-kube-api-access-x7752\") pod \"sg-bridge-2-build\" (UID: \"a016ce4f-2506-4886-b6d2-3a51ce64dab0\") " pod="service-telemetry/sg-bridge-2-build" Oct 08 00:25:51 crc kubenswrapper[4967]: I1008 00:25:51.403885 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/a016ce4f-2506-4886-b6d2-3a51ce64dab0-container-storage-root\") pod \"sg-bridge-2-build\" (UID: \"a016ce4f-2506-4886-b6d2-3a51ce64dab0\") " pod="service-telemetry/sg-bridge-2-build" Oct 08 00:25:51 crc kubenswrapper[4967]: I1008 00:25:51.404019 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a016ce4f-2506-4886-b6d2-3a51ce64dab0-build-ca-bundles\") pod \"sg-bridge-2-build\" (UID: \"a016ce4f-2506-4886-b6d2-3a51ce64dab0\") " pod="service-telemetry/sg-bridge-2-build" Oct 08 00:25:51 crc kubenswrapper[4967]: I1008 00:25:51.404089 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/a016ce4f-2506-4886-b6d2-3a51ce64dab0-container-storage-run\") pod \"sg-bridge-2-build\" (UID: \"a016ce4f-2506-4886-b6d2-3a51ce64dab0\") " pod="service-telemetry/sg-bridge-2-build" Oct 08 00:25:51 crc kubenswrapper[4967]: I1008 00:25:51.404153 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-r7snr-push\" (UniqueName: \"kubernetes.io/secret/a016ce4f-2506-4886-b6d2-3a51ce64dab0-builder-dockercfg-r7snr-push\") pod \"sg-bridge-2-build\" (UID: \"a016ce4f-2506-4886-b6d2-3a51ce64dab0\") " pod="service-telemetry/sg-bridge-2-build" Oct 08 00:25:51 crc kubenswrapper[4967]: I1008 00:25:51.404201 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/a016ce4f-2506-4886-b6d2-3a51ce64dab0-buildcachedir\") pod \"sg-bridge-2-build\" (UID: \"a016ce4f-2506-4886-b6d2-3a51ce64dab0\") " pod="service-telemetry/sg-bridge-2-build" Oct 08 00:25:51 crc kubenswrapper[4967]: I1008 00:25:51.505521 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/a016ce4f-2506-4886-b6d2-3a51ce64dab0-build-system-configs\") pod \"sg-bridge-2-build\" (UID: \"a016ce4f-2506-4886-b6d2-3a51ce64dab0\") " pod="service-telemetry/sg-bridge-2-build" Oct 08 00:25:51 crc kubenswrapper[4967]: I1008 00:25:51.505599 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/a016ce4f-2506-4886-b6d2-3a51ce64dab0-buildworkdir\") pod \"sg-bridge-2-build\" (UID: \"a016ce4f-2506-4886-b6d2-3a51ce64dab0\") " pod="service-telemetry/sg-bridge-2-build" Oct 08 00:25:51 crc kubenswrapper[4967]: I1008 00:25:51.505649 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/a016ce4f-2506-4886-b6d2-3a51ce64dab0-node-pullsecrets\") pod \"sg-bridge-2-build\" (UID: \"a016ce4f-2506-4886-b6d2-3a51ce64dab0\") " pod="service-telemetry/sg-bridge-2-build" Oct 08 00:25:51 crc kubenswrapper[4967]: I1008 00:25:51.505770 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x7752\" (UniqueName: \"kubernetes.io/projected/a016ce4f-2506-4886-b6d2-3a51ce64dab0-kube-api-access-x7752\") pod \"sg-bridge-2-build\" (UID: \"a016ce4f-2506-4886-b6d2-3a51ce64dab0\") " pod="service-telemetry/sg-bridge-2-build" Oct 08 00:25:51 crc kubenswrapper[4967]: I1008 00:25:51.505812 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/a016ce4f-2506-4886-b6d2-3a51ce64dab0-container-storage-root\") pod \"sg-bridge-2-build\" (UID: \"a016ce4f-2506-4886-b6d2-3a51ce64dab0\") " pod="service-telemetry/sg-bridge-2-build" Oct 08 00:25:51 crc kubenswrapper[4967]: I1008 00:25:51.505876 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a016ce4f-2506-4886-b6d2-3a51ce64dab0-build-ca-bundles\") pod \"sg-bridge-2-build\" (UID: \"a016ce4f-2506-4886-b6d2-3a51ce64dab0\") " pod="service-telemetry/sg-bridge-2-build" Oct 08 00:25:51 crc kubenswrapper[4967]: I1008 00:25:51.505914 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/a016ce4f-2506-4886-b6d2-3a51ce64dab0-container-storage-run\") pod \"sg-bridge-2-build\" (UID: \"a016ce4f-2506-4886-b6d2-3a51ce64dab0\") " pod="service-telemetry/sg-bridge-2-build" Oct 08 00:25:51 crc kubenswrapper[4967]: I1008 00:25:51.505919 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/a016ce4f-2506-4886-b6d2-3a51ce64dab0-node-pullsecrets\") pod \"sg-bridge-2-build\" (UID: \"a016ce4f-2506-4886-b6d2-3a51ce64dab0\") " pod="service-telemetry/sg-bridge-2-build" Oct 08 00:25:51 crc kubenswrapper[4967]: I1008 00:25:51.505952 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-r7snr-push\" (UniqueName: \"kubernetes.io/secret/a016ce4f-2506-4886-b6d2-3a51ce64dab0-builder-dockercfg-r7snr-push\") pod \"sg-bridge-2-build\" (UID: \"a016ce4f-2506-4886-b6d2-3a51ce64dab0\") " pod="service-telemetry/sg-bridge-2-build" Oct 08 00:25:51 crc kubenswrapper[4967]: I1008 00:25:51.506054 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/a016ce4f-2506-4886-b6d2-3a51ce64dab0-buildcachedir\") pod \"sg-bridge-2-build\" (UID: \"a016ce4f-2506-4886-b6d2-3a51ce64dab0\") " pod="service-telemetry/sg-bridge-2-build" Oct 08 00:25:51 crc kubenswrapper[4967]: I1008 00:25:51.506092 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-r7snr-pull\" (UniqueName: \"kubernetes.io/secret/a016ce4f-2506-4886-b6d2-3a51ce64dab0-builder-dockercfg-r7snr-pull\") pod \"sg-bridge-2-build\" (UID: \"a016ce4f-2506-4886-b6d2-3a51ce64dab0\") " pod="service-telemetry/sg-bridge-2-build" Oct 08 00:25:51 crc kubenswrapper[4967]: I1008 00:25:51.506114 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a016ce4f-2506-4886-b6d2-3a51ce64dab0-build-proxy-ca-bundles\") pod \"sg-bridge-2-build\" (UID: \"a016ce4f-2506-4886-b6d2-3a51ce64dab0\") " pod="service-telemetry/sg-bridge-2-build" Oct 08 00:25:51 crc kubenswrapper[4967]: I1008 00:25:51.506137 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/a016ce4f-2506-4886-b6d2-3a51ce64dab0-build-blob-cache\") pod \"sg-bridge-2-build\" (UID: \"a016ce4f-2506-4886-b6d2-3a51ce64dab0\") " pod="service-telemetry/sg-bridge-2-build" Oct 08 00:25:51 crc kubenswrapper[4967]: I1008 00:25:51.506194 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/a016ce4f-2506-4886-b6d2-3a51ce64dab0-buildworkdir\") pod \"sg-bridge-2-build\" (UID: \"a016ce4f-2506-4886-b6d2-3a51ce64dab0\") " pod="service-telemetry/sg-bridge-2-build" Oct 08 00:25:51 crc kubenswrapper[4967]: I1008 00:25:51.506637 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/a016ce4f-2506-4886-b6d2-3a51ce64dab0-build-blob-cache\") pod \"sg-bridge-2-build\" (UID: \"a016ce4f-2506-4886-b6d2-3a51ce64dab0\") " pod="service-telemetry/sg-bridge-2-build" Oct 08 00:25:51 crc kubenswrapper[4967]: I1008 00:25:51.506827 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/a016ce4f-2506-4886-b6d2-3a51ce64dab0-build-system-configs\") pod \"sg-bridge-2-build\" (UID: \"a016ce4f-2506-4886-b6d2-3a51ce64dab0\") " pod="service-telemetry/sg-bridge-2-build" Oct 08 00:25:51 crc kubenswrapper[4967]: I1008 00:25:51.506841 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/a016ce4f-2506-4886-b6d2-3a51ce64dab0-buildcachedir\") pod \"sg-bridge-2-build\" (UID: \"a016ce4f-2506-4886-b6d2-3a51ce64dab0\") " pod="service-telemetry/sg-bridge-2-build" Oct 08 00:25:51 crc kubenswrapper[4967]: I1008 00:25:51.506898 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a016ce4f-2506-4886-b6d2-3a51ce64dab0-build-proxy-ca-bundles\") pod \"sg-bridge-2-build\" (UID: \"a016ce4f-2506-4886-b6d2-3a51ce64dab0\") " pod="service-telemetry/sg-bridge-2-build" Oct 08 00:25:51 crc kubenswrapper[4967]: I1008 00:25:51.507298 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/a016ce4f-2506-4886-b6d2-3a51ce64dab0-container-storage-root\") pod \"sg-bridge-2-build\" (UID: \"a016ce4f-2506-4886-b6d2-3a51ce64dab0\") " pod="service-telemetry/sg-bridge-2-build" Oct 08 00:25:51 crc kubenswrapper[4967]: I1008 00:25:51.507431 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/a016ce4f-2506-4886-b6d2-3a51ce64dab0-container-storage-run\") pod \"sg-bridge-2-build\" (UID: \"a016ce4f-2506-4886-b6d2-3a51ce64dab0\") " pod="service-telemetry/sg-bridge-2-build" Oct 08 00:25:51 crc kubenswrapper[4967]: I1008 00:25:51.507574 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a016ce4f-2506-4886-b6d2-3a51ce64dab0-build-ca-bundles\") pod \"sg-bridge-2-build\" (UID: \"a016ce4f-2506-4886-b6d2-3a51ce64dab0\") " pod="service-telemetry/sg-bridge-2-build" Oct 08 00:25:51 crc kubenswrapper[4967]: I1008 00:25:51.510102 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-r7snr-push\" (UniqueName: \"kubernetes.io/secret/a016ce4f-2506-4886-b6d2-3a51ce64dab0-builder-dockercfg-r7snr-push\") pod \"sg-bridge-2-build\" (UID: \"a016ce4f-2506-4886-b6d2-3a51ce64dab0\") " pod="service-telemetry/sg-bridge-2-build" Oct 08 00:25:51 crc kubenswrapper[4967]: I1008 00:25:51.510122 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-r7snr-pull\" (UniqueName: \"kubernetes.io/secret/a016ce4f-2506-4886-b6d2-3a51ce64dab0-builder-dockercfg-r7snr-pull\") pod \"sg-bridge-2-build\" (UID: \"a016ce4f-2506-4886-b6d2-3a51ce64dab0\") " pod="service-telemetry/sg-bridge-2-build" Oct 08 00:25:51 crc kubenswrapper[4967]: I1008 00:25:51.528253 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x7752\" (UniqueName: \"kubernetes.io/projected/a016ce4f-2506-4886-b6d2-3a51ce64dab0-kube-api-access-x7752\") pod \"sg-bridge-2-build\" (UID: \"a016ce4f-2506-4886-b6d2-3a51ce64dab0\") " pod="service-telemetry/sg-bridge-2-build" Oct 08 00:25:51 crc kubenswrapper[4967]: I1008 00:25:51.672106 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-bridge-2-build" Oct 08 00:25:51 crc kubenswrapper[4967]: I1008 00:25:51.914001 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/sg-bridge-2-build"] Oct 08 00:25:52 crc kubenswrapper[4967]: I1008 00:25:52.006099 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-2-build" event={"ID":"a016ce4f-2506-4886-b6d2-3a51ce64dab0","Type":"ContainerStarted","Data":"798e1146e395a5fd0eae984499f1e835f80a9af76c932608e26cbf50ccdc4d58"} Oct 08 00:25:52 crc kubenswrapper[4967]: I1008 00:25:52.130198 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6cc52f44-a91c-45d1-b86a-7e658bf4cd5b" path="/var/lib/kubelet/pods/6cc52f44-a91c-45d1-b86a-7e658bf4cd5b/volumes" Oct 08 00:25:53 crc kubenswrapper[4967]: I1008 00:25:53.017353 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-2-build" event={"ID":"a016ce4f-2506-4886-b6d2-3a51ce64dab0","Type":"ContainerStarted","Data":"57f02743020433a5f4d97ee9f6a6a2fac79d3e1122ae4aca764cf2d6786f330b"} Oct 08 00:25:54 crc kubenswrapper[4967]: I1008 00:25:54.026939 4967 generic.go:334] "Generic (PLEG): container finished" podID="a016ce4f-2506-4886-b6d2-3a51ce64dab0" containerID="57f02743020433a5f4d97ee9f6a6a2fac79d3e1122ae4aca764cf2d6786f330b" exitCode=0 Oct 08 00:25:54 crc kubenswrapper[4967]: I1008 00:25:54.027079 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-2-build" event={"ID":"a016ce4f-2506-4886-b6d2-3a51ce64dab0","Type":"ContainerDied","Data":"57f02743020433a5f4d97ee9f6a6a2fac79d3e1122ae4aca764cf2d6786f330b"} Oct 08 00:25:55 crc kubenswrapper[4967]: I1008 00:25:55.035998 4967 generic.go:334] "Generic (PLEG): container finished" podID="a016ce4f-2506-4886-b6d2-3a51ce64dab0" containerID="d64f847735d2d07822f91f1d556cc4d19f71e4b4c09f88064f1fee16f1dbe4ba" exitCode=0 Oct 08 00:25:55 crc kubenswrapper[4967]: I1008 00:25:55.036068 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-2-build" event={"ID":"a016ce4f-2506-4886-b6d2-3a51ce64dab0","Type":"ContainerDied","Data":"d64f847735d2d07822f91f1d556cc4d19f71e4b4c09f88064f1fee16f1dbe4ba"} Oct 08 00:25:55 crc kubenswrapper[4967]: I1008 00:25:55.086068 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_sg-bridge-2-build_a016ce4f-2506-4886-b6d2-3a51ce64dab0/manage-dockerfile/0.log" Oct 08 00:25:56 crc kubenswrapper[4967]: I1008 00:25:56.048543 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-2-build" event={"ID":"a016ce4f-2506-4886-b6d2-3a51ce64dab0","Type":"ContainerStarted","Data":"3a01a167520df7d4d93d9443f1446a9f51823c8a1b98b1bb881d106837414907"} Oct 08 00:25:56 crc kubenswrapper[4967]: I1008 00:25:56.108388 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/sg-bridge-2-build" podStartSLOduration=5.108363556 podStartE2EDuration="5.108363556s" podCreationTimestamp="2025-10-08 00:25:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:25:56.102001025 +0000 UTC m=+1154.752535694" watchObservedRunningTime="2025-10-08 00:25:56.108363556 +0000 UTC m=+1154.758898215" Oct 08 00:26:36 crc kubenswrapper[4967]: I1008 00:26:36.920557 4967 patch_prober.go:28] interesting pod/machine-config-daemon-xvdtp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 00:26:36 crc kubenswrapper[4967]: I1008 00:26:36.921151 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" podUID="9461162a-c27d-465e-9ca7-2d3936877cd8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 00:26:42 crc kubenswrapper[4967]: I1008 00:26:42.426258 4967 generic.go:334] "Generic (PLEG): container finished" podID="a016ce4f-2506-4886-b6d2-3a51ce64dab0" containerID="3a01a167520df7d4d93d9443f1446a9f51823c8a1b98b1bb881d106837414907" exitCode=0 Oct 08 00:26:42 crc kubenswrapper[4967]: I1008 00:26:42.426404 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-2-build" event={"ID":"a016ce4f-2506-4886-b6d2-3a51ce64dab0","Type":"ContainerDied","Data":"3a01a167520df7d4d93d9443f1446a9f51823c8a1b98b1bb881d106837414907"} Oct 08 00:26:43 crc kubenswrapper[4967]: I1008 00:26:43.781340 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-bridge-2-build" Oct 08 00:26:43 crc kubenswrapper[4967]: I1008 00:26:43.898386 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/a016ce4f-2506-4886-b6d2-3a51ce64dab0-container-storage-run\") pod \"a016ce4f-2506-4886-b6d2-3a51ce64dab0\" (UID: \"a016ce4f-2506-4886-b6d2-3a51ce64dab0\") " Oct 08 00:26:43 crc kubenswrapper[4967]: I1008 00:26:43.898489 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-r7snr-push\" (UniqueName: \"kubernetes.io/secret/a016ce4f-2506-4886-b6d2-3a51ce64dab0-builder-dockercfg-r7snr-push\") pod \"a016ce4f-2506-4886-b6d2-3a51ce64dab0\" (UID: \"a016ce4f-2506-4886-b6d2-3a51ce64dab0\") " Oct 08 00:26:43 crc kubenswrapper[4967]: I1008 00:26:43.898530 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/a016ce4f-2506-4886-b6d2-3a51ce64dab0-build-blob-cache\") pod \"a016ce4f-2506-4886-b6d2-3a51ce64dab0\" (UID: \"a016ce4f-2506-4886-b6d2-3a51ce64dab0\") " Oct 08 00:26:43 crc kubenswrapper[4967]: I1008 00:26:43.898573 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/a016ce4f-2506-4886-b6d2-3a51ce64dab0-node-pullsecrets\") pod \"a016ce4f-2506-4886-b6d2-3a51ce64dab0\" (UID: \"a016ce4f-2506-4886-b6d2-3a51ce64dab0\") " Oct 08 00:26:43 crc kubenswrapper[4967]: I1008 00:26:43.898613 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a016ce4f-2506-4886-b6d2-3a51ce64dab0-build-proxy-ca-bundles\") pod \"a016ce4f-2506-4886-b6d2-3a51ce64dab0\" (UID: \"a016ce4f-2506-4886-b6d2-3a51ce64dab0\") " Oct 08 00:26:43 crc kubenswrapper[4967]: I1008 00:26:43.898735 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/a016ce4f-2506-4886-b6d2-3a51ce64dab0-container-storage-root\") pod \"a016ce4f-2506-4886-b6d2-3a51ce64dab0\" (UID: \"a016ce4f-2506-4886-b6d2-3a51ce64dab0\") " Oct 08 00:26:43 crc kubenswrapper[4967]: I1008 00:26:43.898763 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/a016ce4f-2506-4886-b6d2-3a51ce64dab0-buildcachedir\") pod \"a016ce4f-2506-4886-b6d2-3a51ce64dab0\" (UID: \"a016ce4f-2506-4886-b6d2-3a51ce64dab0\") " Oct 08 00:26:43 crc kubenswrapper[4967]: I1008 00:26:43.898798 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/a016ce4f-2506-4886-b6d2-3a51ce64dab0-buildworkdir\") pod \"a016ce4f-2506-4886-b6d2-3a51ce64dab0\" (UID: \"a016ce4f-2506-4886-b6d2-3a51ce64dab0\") " Oct 08 00:26:43 crc kubenswrapper[4967]: I1008 00:26:43.898840 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7752\" (UniqueName: \"kubernetes.io/projected/a016ce4f-2506-4886-b6d2-3a51ce64dab0-kube-api-access-x7752\") pod \"a016ce4f-2506-4886-b6d2-3a51ce64dab0\" (UID: \"a016ce4f-2506-4886-b6d2-3a51ce64dab0\") " Oct 08 00:26:43 crc kubenswrapper[4967]: I1008 00:26:43.898894 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-r7snr-pull\" (UniqueName: \"kubernetes.io/secret/a016ce4f-2506-4886-b6d2-3a51ce64dab0-builder-dockercfg-r7snr-pull\") pod \"a016ce4f-2506-4886-b6d2-3a51ce64dab0\" (UID: \"a016ce4f-2506-4886-b6d2-3a51ce64dab0\") " Oct 08 00:26:43 crc kubenswrapper[4967]: I1008 00:26:43.898955 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a016ce4f-2506-4886-b6d2-3a51ce64dab0-build-ca-bundles\") pod \"a016ce4f-2506-4886-b6d2-3a51ce64dab0\" (UID: \"a016ce4f-2506-4886-b6d2-3a51ce64dab0\") " Oct 08 00:26:43 crc kubenswrapper[4967]: I1008 00:26:43.899003 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/a016ce4f-2506-4886-b6d2-3a51ce64dab0-build-system-configs\") pod \"a016ce4f-2506-4886-b6d2-3a51ce64dab0\" (UID: \"a016ce4f-2506-4886-b6d2-3a51ce64dab0\") " Oct 08 00:26:43 crc kubenswrapper[4967]: I1008 00:26:43.899831 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a016ce4f-2506-4886-b6d2-3a51ce64dab0-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "a016ce4f-2506-4886-b6d2-3a51ce64dab0" (UID: "a016ce4f-2506-4886-b6d2-3a51ce64dab0"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 00:26:43 crc kubenswrapper[4967]: I1008 00:26:43.899917 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a016ce4f-2506-4886-b6d2-3a51ce64dab0-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "a016ce4f-2506-4886-b6d2-3a51ce64dab0" (UID: "a016ce4f-2506-4886-b6d2-3a51ce64dab0"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 00:26:43 crc kubenswrapper[4967]: I1008 00:26:43.901302 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a016ce4f-2506-4886-b6d2-3a51ce64dab0-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "a016ce4f-2506-4886-b6d2-3a51ce64dab0" (UID: "a016ce4f-2506-4886-b6d2-3a51ce64dab0"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:26:43 crc kubenswrapper[4967]: I1008 00:26:43.901485 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a016ce4f-2506-4886-b6d2-3a51ce64dab0-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "a016ce4f-2506-4886-b6d2-3a51ce64dab0" (UID: "a016ce4f-2506-4886-b6d2-3a51ce64dab0"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:26:43 crc kubenswrapper[4967]: I1008 00:26:43.903449 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a016ce4f-2506-4886-b6d2-3a51ce64dab0-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "a016ce4f-2506-4886-b6d2-3a51ce64dab0" (UID: "a016ce4f-2506-4886-b6d2-3a51ce64dab0"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:26:43 crc kubenswrapper[4967]: I1008 00:26:43.904472 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a016ce4f-2506-4886-b6d2-3a51ce64dab0-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "a016ce4f-2506-4886-b6d2-3a51ce64dab0" (UID: "a016ce4f-2506-4886-b6d2-3a51ce64dab0"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:26:43 crc kubenswrapper[4967]: I1008 00:26:43.907544 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a016ce4f-2506-4886-b6d2-3a51ce64dab0-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "a016ce4f-2506-4886-b6d2-3a51ce64dab0" (UID: "a016ce4f-2506-4886-b6d2-3a51ce64dab0"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:26:43 crc kubenswrapper[4967]: I1008 00:26:43.908345 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a016ce4f-2506-4886-b6d2-3a51ce64dab0-builder-dockercfg-r7snr-push" (OuterVolumeSpecName: "builder-dockercfg-r7snr-push") pod "a016ce4f-2506-4886-b6d2-3a51ce64dab0" (UID: "a016ce4f-2506-4886-b6d2-3a51ce64dab0"). InnerVolumeSpecName "builder-dockercfg-r7snr-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:26:43 crc kubenswrapper[4967]: I1008 00:26:43.909168 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a016ce4f-2506-4886-b6d2-3a51ce64dab0-kube-api-access-x7752" (OuterVolumeSpecName: "kube-api-access-x7752") pod "a016ce4f-2506-4886-b6d2-3a51ce64dab0" (UID: "a016ce4f-2506-4886-b6d2-3a51ce64dab0"). InnerVolumeSpecName "kube-api-access-x7752". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:26:43 crc kubenswrapper[4967]: I1008 00:26:43.909757 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a016ce4f-2506-4886-b6d2-3a51ce64dab0-builder-dockercfg-r7snr-pull" (OuterVolumeSpecName: "builder-dockercfg-r7snr-pull") pod "a016ce4f-2506-4886-b6d2-3a51ce64dab0" (UID: "a016ce4f-2506-4886-b6d2-3a51ce64dab0"). InnerVolumeSpecName "builder-dockercfg-r7snr-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:26:44 crc kubenswrapper[4967]: I1008 00:26:44.000517 4967 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/a016ce4f-2506-4886-b6d2-3a51ce64dab0-build-system-configs\") on node \"crc\" DevicePath \"\"" Oct 08 00:26:44 crc kubenswrapper[4967]: I1008 00:26:44.000568 4967 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/a016ce4f-2506-4886-b6d2-3a51ce64dab0-container-storage-run\") on node \"crc\" DevicePath \"\"" Oct 08 00:26:44 crc kubenswrapper[4967]: I1008 00:26:44.000590 4967 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-r7snr-push\" (UniqueName: \"kubernetes.io/secret/a016ce4f-2506-4886-b6d2-3a51ce64dab0-builder-dockercfg-r7snr-push\") on node \"crc\" DevicePath \"\"" Oct 08 00:26:44 crc kubenswrapper[4967]: I1008 00:26:44.000610 4967 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/a016ce4f-2506-4886-b6d2-3a51ce64dab0-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Oct 08 00:26:44 crc kubenswrapper[4967]: I1008 00:26:44.000628 4967 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a016ce4f-2506-4886-b6d2-3a51ce64dab0-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 08 00:26:44 crc kubenswrapper[4967]: I1008 00:26:44.000646 4967 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/a016ce4f-2506-4886-b6d2-3a51ce64dab0-buildcachedir\") on node \"crc\" DevicePath \"\"" Oct 08 00:26:44 crc kubenswrapper[4967]: I1008 00:26:44.000664 4967 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/a016ce4f-2506-4886-b6d2-3a51ce64dab0-buildworkdir\") on node \"crc\" DevicePath \"\"" Oct 08 00:26:44 crc kubenswrapper[4967]: I1008 00:26:44.000706 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7752\" (UniqueName: \"kubernetes.io/projected/a016ce4f-2506-4886-b6d2-3a51ce64dab0-kube-api-access-x7752\") on node \"crc\" DevicePath \"\"" Oct 08 00:26:44 crc kubenswrapper[4967]: I1008 00:26:44.000725 4967 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-r7snr-pull\" (UniqueName: \"kubernetes.io/secret/a016ce4f-2506-4886-b6d2-3a51ce64dab0-builder-dockercfg-r7snr-pull\") on node \"crc\" DevicePath \"\"" Oct 08 00:26:44 crc kubenswrapper[4967]: I1008 00:26:44.000744 4967 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a016ce4f-2506-4886-b6d2-3a51ce64dab0-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 08 00:26:44 crc kubenswrapper[4967]: I1008 00:26:44.080439 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a016ce4f-2506-4886-b6d2-3a51ce64dab0-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "a016ce4f-2506-4886-b6d2-3a51ce64dab0" (UID: "a016ce4f-2506-4886-b6d2-3a51ce64dab0"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:26:44 crc kubenswrapper[4967]: I1008 00:26:44.102118 4967 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/a016ce4f-2506-4886-b6d2-3a51ce64dab0-build-blob-cache\") on node \"crc\" DevicePath \"\"" Oct 08 00:26:44 crc kubenswrapper[4967]: I1008 00:26:44.460417 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/sg-bridge-2-build" Oct 08 00:26:44 crc kubenswrapper[4967]: I1008 00:26:44.461205 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/sg-bridge-2-build" event={"ID":"a016ce4f-2506-4886-b6d2-3a51ce64dab0","Type":"ContainerDied","Data":"798e1146e395a5fd0eae984499f1e835f80a9af76c932608e26cbf50ccdc4d58"} Oct 08 00:26:44 crc kubenswrapper[4967]: I1008 00:26:44.461272 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="798e1146e395a5fd0eae984499f1e835f80a9af76c932608e26cbf50ccdc4d58" Oct 08 00:26:45 crc kubenswrapper[4967]: I1008 00:26:45.064804 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a016ce4f-2506-4886-b6d2-3a51ce64dab0-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "a016ce4f-2506-4886-b6d2-3a51ce64dab0" (UID: "a016ce4f-2506-4886-b6d2-3a51ce64dab0"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:26:45 crc kubenswrapper[4967]: I1008 00:26:45.124582 4967 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/a016ce4f-2506-4886-b6d2-3a51ce64dab0-container-storage-root\") on node \"crc\" DevicePath \"\"" Oct 08 00:26:48 crc kubenswrapper[4967]: I1008 00:26:48.689358 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/prometheus-webhook-snmp-1-build"] Oct 08 00:26:48 crc kubenswrapper[4967]: E1008 00:26:48.690006 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a016ce4f-2506-4886-b6d2-3a51ce64dab0" containerName="git-clone" Oct 08 00:26:48 crc kubenswrapper[4967]: I1008 00:26:48.690028 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="a016ce4f-2506-4886-b6d2-3a51ce64dab0" containerName="git-clone" Oct 08 00:26:48 crc kubenswrapper[4967]: E1008 00:26:48.690061 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a016ce4f-2506-4886-b6d2-3a51ce64dab0" containerName="docker-build" Oct 08 00:26:48 crc kubenswrapper[4967]: I1008 00:26:48.690075 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="a016ce4f-2506-4886-b6d2-3a51ce64dab0" containerName="docker-build" Oct 08 00:26:48 crc kubenswrapper[4967]: E1008 00:26:48.690105 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a016ce4f-2506-4886-b6d2-3a51ce64dab0" containerName="manage-dockerfile" Oct 08 00:26:48 crc kubenswrapper[4967]: I1008 00:26:48.690122 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="a016ce4f-2506-4886-b6d2-3a51ce64dab0" containerName="manage-dockerfile" Oct 08 00:26:48 crc kubenswrapper[4967]: I1008 00:26:48.692761 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="a016ce4f-2506-4886-b6d2-3a51ce64dab0" containerName="docker-build" Oct 08 00:26:48 crc kubenswrapper[4967]: I1008 00:26:48.693913 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-webhook-snmp-1-build" Oct 08 00:26:48 crc kubenswrapper[4967]: I1008 00:26:48.695825 4967 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-r7snr" Oct 08 00:26:48 crc kubenswrapper[4967]: I1008 00:26:48.696169 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"prometheus-webhook-snmp-1-ca" Oct 08 00:26:48 crc kubenswrapper[4967]: I1008 00:26:48.702485 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"prometheus-webhook-snmp-1-sys-config" Oct 08 00:26:48 crc kubenswrapper[4967]: I1008 00:26:48.702612 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"prometheus-webhook-snmp-1-global-ca" Oct 08 00:26:48 crc kubenswrapper[4967]: I1008 00:26:48.711983 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/prometheus-webhook-snmp-1-build"] Oct 08 00:26:48 crc kubenswrapper[4967]: I1008 00:26:48.879477 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/b742fb05-bba8-4bc5-8ac7-97e1061a8842-container-storage-root\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"b742fb05-bba8-4bc5-8ac7-97e1061a8842\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Oct 08 00:26:48 crc kubenswrapper[4967]: I1008 00:26:48.879610 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/b742fb05-bba8-4bc5-8ac7-97e1061a8842-build-blob-cache\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"b742fb05-bba8-4bc5-8ac7-97e1061a8842\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Oct 08 00:26:48 crc kubenswrapper[4967]: I1008 00:26:48.879791 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-r7snr-pull\" (UniqueName: \"kubernetes.io/secret/b742fb05-bba8-4bc5-8ac7-97e1061a8842-builder-dockercfg-r7snr-pull\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"b742fb05-bba8-4bc5-8ac7-97e1061a8842\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Oct 08 00:26:48 crc kubenswrapper[4967]: I1008 00:26:48.879835 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rwjzq\" (UniqueName: \"kubernetes.io/projected/b742fb05-bba8-4bc5-8ac7-97e1061a8842-kube-api-access-rwjzq\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"b742fb05-bba8-4bc5-8ac7-97e1061a8842\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Oct 08 00:26:48 crc kubenswrapper[4967]: I1008 00:26:48.879910 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b742fb05-bba8-4bc5-8ac7-97e1061a8842-build-proxy-ca-bundles\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"b742fb05-bba8-4bc5-8ac7-97e1061a8842\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Oct 08 00:26:48 crc kubenswrapper[4967]: I1008 00:26:48.879953 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/b742fb05-bba8-4bc5-8ac7-97e1061a8842-buildcachedir\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"b742fb05-bba8-4bc5-8ac7-97e1061a8842\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Oct 08 00:26:48 crc kubenswrapper[4967]: I1008 00:26:48.880013 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/b742fb05-bba8-4bc5-8ac7-97e1061a8842-build-system-configs\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"b742fb05-bba8-4bc5-8ac7-97e1061a8842\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Oct 08 00:26:48 crc kubenswrapper[4967]: I1008 00:26:48.880060 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b742fb05-bba8-4bc5-8ac7-97e1061a8842-build-ca-bundles\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"b742fb05-bba8-4bc5-8ac7-97e1061a8842\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Oct 08 00:26:48 crc kubenswrapper[4967]: I1008 00:26:48.880093 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/b742fb05-bba8-4bc5-8ac7-97e1061a8842-container-storage-run\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"b742fb05-bba8-4bc5-8ac7-97e1061a8842\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Oct 08 00:26:48 crc kubenswrapper[4967]: I1008 00:26:48.880203 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/b742fb05-bba8-4bc5-8ac7-97e1061a8842-node-pullsecrets\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"b742fb05-bba8-4bc5-8ac7-97e1061a8842\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Oct 08 00:26:48 crc kubenswrapper[4967]: I1008 00:26:48.880238 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/b742fb05-bba8-4bc5-8ac7-97e1061a8842-buildworkdir\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"b742fb05-bba8-4bc5-8ac7-97e1061a8842\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Oct 08 00:26:48 crc kubenswrapper[4967]: I1008 00:26:48.880296 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-r7snr-push\" (UniqueName: \"kubernetes.io/secret/b742fb05-bba8-4bc5-8ac7-97e1061a8842-builder-dockercfg-r7snr-push\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"b742fb05-bba8-4bc5-8ac7-97e1061a8842\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Oct 08 00:26:48 crc kubenswrapper[4967]: I1008 00:26:48.990979 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b742fb05-bba8-4bc5-8ac7-97e1061a8842-build-proxy-ca-bundles\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"b742fb05-bba8-4bc5-8ac7-97e1061a8842\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Oct 08 00:26:48 crc kubenswrapper[4967]: I1008 00:26:48.991104 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/b742fb05-bba8-4bc5-8ac7-97e1061a8842-buildcachedir\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"b742fb05-bba8-4bc5-8ac7-97e1061a8842\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Oct 08 00:26:48 crc kubenswrapper[4967]: I1008 00:26:48.991291 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/b742fb05-bba8-4bc5-8ac7-97e1061a8842-build-system-configs\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"b742fb05-bba8-4bc5-8ac7-97e1061a8842\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Oct 08 00:26:48 crc kubenswrapper[4967]: I1008 00:26:48.991458 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/b742fb05-bba8-4bc5-8ac7-97e1061a8842-buildcachedir\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"b742fb05-bba8-4bc5-8ac7-97e1061a8842\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Oct 08 00:26:48 crc kubenswrapper[4967]: I1008 00:26:48.991478 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b742fb05-bba8-4bc5-8ac7-97e1061a8842-build-ca-bundles\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"b742fb05-bba8-4bc5-8ac7-97e1061a8842\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Oct 08 00:26:48 crc kubenswrapper[4967]: I1008 00:26:48.991719 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/b742fb05-bba8-4bc5-8ac7-97e1061a8842-container-storage-run\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"b742fb05-bba8-4bc5-8ac7-97e1061a8842\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Oct 08 00:26:48 crc kubenswrapper[4967]: I1008 00:26:48.991855 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/b742fb05-bba8-4bc5-8ac7-97e1061a8842-node-pullsecrets\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"b742fb05-bba8-4bc5-8ac7-97e1061a8842\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Oct 08 00:26:48 crc kubenswrapper[4967]: I1008 00:26:48.992304 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/b742fb05-bba8-4bc5-8ac7-97e1061a8842-buildworkdir\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"b742fb05-bba8-4bc5-8ac7-97e1061a8842\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Oct 08 00:26:48 crc kubenswrapper[4967]: I1008 00:26:48.992488 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-r7snr-push\" (UniqueName: \"kubernetes.io/secret/b742fb05-bba8-4bc5-8ac7-97e1061a8842-builder-dockercfg-r7snr-push\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"b742fb05-bba8-4bc5-8ac7-97e1061a8842\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Oct 08 00:26:48 crc kubenswrapper[4967]: I1008 00:26:48.992603 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/b742fb05-bba8-4bc5-8ac7-97e1061a8842-build-blob-cache\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"b742fb05-bba8-4bc5-8ac7-97e1061a8842\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Oct 08 00:26:48 crc kubenswrapper[4967]: I1008 00:26:48.992699 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/b742fb05-bba8-4bc5-8ac7-97e1061a8842-container-storage-root\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"b742fb05-bba8-4bc5-8ac7-97e1061a8842\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Oct 08 00:26:48 crc kubenswrapper[4967]: I1008 00:26:48.993251 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/b742fb05-bba8-4bc5-8ac7-97e1061a8842-node-pullsecrets\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"b742fb05-bba8-4bc5-8ac7-97e1061a8842\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Oct 08 00:26:48 crc kubenswrapper[4967]: I1008 00:26:48.993808 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/b742fb05-bba8-4bc5-8ac7-97e1061a8842-build-system-configs\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"b742fb05-bba8-4bc5-8ac7-97e1061a8842\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Oct 08 00:26:48 crc kubenswrapper[4967]: I1008 00:26:48.993858 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-r7snr-pull\" (UniqueName: \"kubernetes.io/secret/b742fb05-bba8-4bc5-8ac7-97e1061a8842-builder-dockercfg-r7snr-pull\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"b742fb05-bba8-4bc5-8ac7-97e1061a8842\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Oct 08 00:26:48 crc kubenswrapper[4967]: I1008 00:26:48.993894 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rwjzq\" (UniqueName: \"kubernetes.io/projected/b742fb05-bba8-4bc5-8ac7-97e1061a8842-kube-api-access-rwjzq\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"b742fb05-bba8-4bc5-8ac7-97e1061a8842\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Oct 08 00:26:48 crc kubenswrapper[4967]: I1008 00:26:48.994229 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/b742fb05-bba8-4bc5-8ac7-97e1061a8842-container-storage-run\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"b742fb05-bba8-4bc5-8ac7-97e1061a8842\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Oct 08 00:26:48 crc kubenswrapper[4967]: I1008 00:26:48.994295 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/b742fb05-bba8-4bc5-8ac7-97e1061a8842-buildworkdir\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"b742fb05-bba8-4bc5-8ac7-97e1061a8842\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Oct 08 00:26:48 crc kubenswrapper[4967]: I1008 00:26:48.994438 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b742fb05-bba8-4bc5-8ac7-97e1061a8842-build-proxy-ca-bundles\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"b742fb05-bba8-4bc5-8ac7-97e1061a8842\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Oct 08 00:26:48 crc kubenswrapper[4967]: I1008 00:26:48.994699 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/b742fb05-bba8-4bc5-8ac7-97e1061a8842-container-storage-root\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"b742fb05-bba8-4bc5-8ac7-97e1061a8842\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Oct 08 00:26:49 crc kubenswrapper[4967]: I1008 00:26:49.000364 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b742fb05-bba8-4bc5-8ac7-97e1061a8842-build-ca-bundles\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"b742fb05-bba8-4bc5-8ac7-97e1061a8842\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Oct 08 00:26:49 crc kubenswrapper[4967]: I1008 00:26:49.121857 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/b742fb05-bba8-4bc5-8ac7-97e1061a8842-build-blob-cache\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"b742fb05-bba8-4bc5-8ac7-97e1061a8842\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Oct 08 00:26:49 crc kubenswrapper[4967]: I1008 00:26:49.122693 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-r7snr-push\" (UniqueName: \"kubernetes.io/secret/b742fb05-bba8-4bc5-8ac7-97e1061a8842-builder-dockercfg-r7snr-push\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"b742fb05-bba8-4bc5-8ac7-97e1061a8842\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Oct 08 00:26:49 crc kubenswrapper[4967]: I1008 00:26:49.122793 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-r7snr-pull\" (UniqueName: \"kubernetes.io/secret/b742fb05-bba8-4bc5-8ac7-97e1061a8842-builder-dockercfg-r7snr-pull\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"b742fb05-bba8-4bc5-8ac7-97e1061a8842\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Oct 08 00:26:49 crc kubenswrapper[4967]: I1008 00:26:49.129843 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rwjzq\" (UniqueName: \"kubernetes.io/projected/b742fb05-bba8-4bc5-8ac7-97e1061a8842-kube-api-access-rwjzq\") pod \"prometheus-webhook-snmp-1-build\" (UID: \"b742fb05-bba8-4bc5-8ac7-97e1061a8842\") " pod="service-telemetry/prometheus-webhook-snmp-1-build" Oct 08 00:26:49 crc kubenswrapper[4967]: I1008 00:26:49.318046 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-webhook-snmp-1-build" Oct 08 00:26:49 crc kubenswrapper[4967]: I1008 00:26:49.805385 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/prometheus-webhook-snmp-1-build"] Oct 08 00:26:50 crc kubenswrapper[4967]: I1008 00:26:50.508459 4967 generic.go:334] "Generic (PLEG): container finished" podID="b742fb05-bba8-4bc5-8ac7-97e1061a8842" containerID="8f97f93a2bf10211cadfd3b6fc1efd07aa96898fb604a1d7c39e29f803acfb2d" exitCode=0 Oct 08 00:26:50 crc kubenswrapper[4967]: I1008 00:26:50.508804 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-1-build" event={"ID":"b742fb05-bba8-4bc5-8ac7-97e1061a8842","Type":"ContainerDied","Data":"8f97f93a2bf10211cadfd3b6fc1efd07aa96898fb604a1d7c39e29f803acfb2d"} Oct 08 00:26:50 crc kubenswrapper[4967]: I1008 00:26:50.509544 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-1-build" event={"ID":"b742fb05-bba8-4bc5-8ac7-97e1061a8842","Type":"ContainerStarted","Data":"a8be1f435a0efc34f9bf4420d630af8251ce95e51b50c711a041676d581159f0"} Oct 08 00:26:51 crc kubenswrapper[4967]: I1008 00:26:51.523742 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-1-build" event={"ID":"b742fb05-bba8-4bc5-8ac7-97e1061a8842","Type":"ContainerStarted","Data":"0c89eecbd340ba06c854d218288b3c21bda552c853702d29d706dd38c00e64d2"} Oct 08 00:26:51 crc kubenswrapper[4967]: I1008 00:26:51.572066 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/prometheus-webhook-snmp-1-build" podStartSLOduration=3.572040132 podStartE2EDuration="3.572040132s" podCreationTimestamp="2025-10-08 00:26:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:26:51.564198028 +0000 UTC m=+1210.214732707" watchObservedRunningTime="2025-10-08 00:26:51.572040132 +0000 UTC m=+1210.222574811" Oct 08 00:26:59 crc kubenswrapper[4967]: I1008 00:26:59.301739 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/prometheus-webhook-snmp-1-build"] Oct 08 00:26:59 crc kubenswrapper[4967]: I1008 00:26:59.302435 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="service-telemetry/prometheus-webhook-snmp-1-build" podUID="b742fb05-bba8-4bc5-8ac7-97e1061a8842" containerName="docker-build" containerID="cri-o://0c89eecbd340ba06c854d218288b3c21bda552c853702d29d706dd38c00e64d2" gracePeriod=30 Oct 08 00:26:59 crc kubenswrapper[4967]: I1008 00:26:59.613623 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_prometheus-webhook-snmp-1-build_b742fb05-bba8-4bc5-8ac7-97e1061a8842/docker-build/0.log" Oct 08 00:26:59 crc kubenswrapper[4967]: I1008 00:26:59.614163 4967 generic.go:334] "Generic (PLEG): container finished" podID="b742fb05-bba8-4bc5-8ac7-97e1061a8842" containerID="0c89eecbd340ba06c854d218288b3c21bda552c853702d29d706dd38c00e64d2" exitCode=1 Oct 08 00:26:59 crc kubenswrapper[4967]: I1008 00:26:59.614197 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-1-build" event={"ID":"b742fb05-bba8-4bc5-8ac7-97e1061a8842","Type":"ContainerDied","Data":"0c89eecbd340ba06c854d218288b3c21bda552c853702d29d706dd38c00e64d2"} Oct 08 00:26:59 crc kubenswrapper[4967]: I1008 00:26:59.773539 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_prometheus-webhook-snmp-1-build_b742fb05-bba8-4bc5-8ac7-97e1061a8842/docker-build/0.log" Oct 08 00:26:59 crc kubenswrapper[4967]: I1008 00:26:59.774421 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-webhook-snmp-1-build" Oct 08 00:26:59 crc kubenswrapper[4967]: I1008 00:26:59.844602 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/b742fb05-bba8-4bc5-8ac7-97e1061a8842-container-storage-root\") pod \"b742fb05-bba8-4bc5-8ac7-97e1061a8842\" (UID: \"b742fb05-bba8-4bc5-8ac7-97e1061a8842\") " Oct 08 00:26:59 crc kubenswrapper[4967]: I1008 00:26:59.844761 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/b742fb05-bba8-4bc5-8ac7-97e1061a8842-buildworkdir\") pod \"b742fb05-bba8-4bc5-8ac7-97e1061a8842\" (UID: \"b742fb05-bba8-4bc5-8ac7-97e1061a8842\") " Oct 08 00:26:59 crc kubenswrapper[4967]: I1008 00:26:59.844839 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b742fb05-bba8-4bc5-8ac7-97e1061a8842-build-ca-bundles\") pod \"b742fb05-bba8-4bc5-8ac7-97e1061a8842\" (UID: \"b742fb05-bba8-4bc5-8ac7-97e1061a8842\") " Oct 08 00:26:59 crc kubenswrapper[4967]: I1008 00:26:59.844882 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-r7snr-push\" (UniqueName: \"kubernetes.io/secret/b742fb05-bba8-4bc5-8ac7-97e1061a8842-builder-dockercfg-r7snr-push\") pod \"b742fb05-bba8-4bc5-8ac7-97e1061a8842\" (UID: \"b742fb05-bba8-4bc5-8ac7-97e1061a8842\") " Oct 08 00:26:59 crc kubenswrapper[4967]: I1008 00:26:59.844924 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/b742fb05-bba8-4bc5-8ac7-97e1061a8842-buildcachedir\") pod \"b742fb05-bba8-4bc5-8ac7-97e1061a8842\" (UID: \"b742fb05-bba8-4bc5-8ac7-97e1061a8842\") " Oct 08 00:26:59 crc kubenswrapper[4967]: I1008 00:26:59.844957 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rwjzq\" (UniqueName: \"kubernetes.io/projected/b742fb05-bba8-4bc5-8ac7-97e1061a8842-kube-api-access-rwjzq\") pod \"b742fb05-bba8-4bc5-8ac7-97e1061a8842\" (UID: \"b742fb05-bba8-4bc5-8ac7-97e1061a8842\") " Oct 08 00:26:59 crc kubenswrapper[4967]: I1008 00:26:59.845018 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/b742fb05-bba8-4bc5-8ac7-97e1061a8842-build-blob-cache\") pod \"b742fb05-bba8-4bc5-8ac7-97e1061a8842\" (UID: \"b742fb05-bba8-4bc5-8ac7-97e1061a8842\") " Oct 08 00:26:59 crc kubenswrapper[4967]: I1008 00:26:59.845050 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/b742fb05-bba8-4bc5-8ac7-97e1061a8842-container-storage-run\") pod \"b742fb05-bba8-4bc5-8ac7-97e1061a8842\" (UID: \"b742fb05-bba8-4bc5-8ac7-97e1061a8842\") " Oct 08 00:26:59 crc kubenswrapper[4967]: I1008 00:26:59.845079 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/b742fb05-bba8-4bc5-8ac7-97e1061a8842-node-pullsecrets\") pod \"b742fb05-bba8-4bc5-8ac7-97e1061a8842\" (UID: \"b742fb05-bba8-4bc5-8ac7-97e1061a8842\") " Oct 08 00:26:59 crc kubenswrapper[4967]: I1008 00:26:59.845120 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b742fb05-bba8-4bc5-8ac7-97e1061a8842-build-proxy-ca-bundles\") pod \"b742fb05-bba8-4bc5-8ac7-97e1061a8842\" (UID: \"b742fb05-bba8-4bc5-8ac7-97e1061a8842\") " Oct 08 00:26:59 crc kubenswrapper[4967]: I1008 00:26:59.845167 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/b742fb05-bba8-4bc5-8ac7-97e1061a8842-build-system-configs\") pod \"b742fb05-bba8-4bc5-8ac7-97e1061a8842\" (UID: \"b742fb05-bba8-4bc5-8ac7-97e1061a8842\") " Oct 08 00:26:59 crc kubenswrapper[4967]: I1008 00:26:59.845284 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-r7snr-pull\" (UniqueName: \"kubernetes.io/secret/b742fb05-bba8-4bc5-8ac7-97e1061a8842-builder-dockercfg-r7snr-pull\") pod \"b742fb05-bba8-4bc5-8ac7-97e1061a8842\" (UID: \"b742fb05-bba8-4bc5-8ac7-97e1061a8842\") " Oct 08 00:26:59 crc kubenswrapper[4967]: I1008 00:26:59.845847 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b742fb05-bba8-4bc5-8ac7-97e1061a8842-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "b742fb05-bba8-4bc5-8ac7-97e1061a8842" (UID: "b742fb05-bba8-4bc5-8ac7-97e1061a8842"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 00:26:59 crc kubenswrapper[4967]: I1008 00:26:59.846322 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b742fb05-bba8-4bc5-8ac7-97e1061a8842-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "b742fb05-bba8-4bc5-8ac7-97e1061a8842" (UID: "b742fb05-bba8-4bc5-8ac7-97e1061a8842"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 00:26:59 crc kubenswrapper[4967]: I1008 00:26:59.846483 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b742fb05-bba8-4bc5-8ac7-97e1061a8842-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "b742fb05-bba8-4bc5-8ac7-97e1061a8842" (UID: "b742fb05-bba8-4bc5-8ac7-97e1061a8842"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:26:59 crc kubenswrapper[4967]: I1008 00:26:59.847073 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b742fb05-bba8-4bc5-8ac7-97e1061a8842-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "b742fb05-bba8-4bc5-8ac7-97e1061a8842" (UID: "b742fb05-bba8-4bc5-8ac7-97e1061a8842"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:26:59 crc kubenswrapper[4967]: I1008 00:26:59.847319 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b742fb05-bba8-4bc5-8ac7-97e1061a8842-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "b742fb05-bba8-4bc5-8ac7-97e1061a8842" (UID: "b742fb05-bba8-4bc5-8ac7-97e1061a8842"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:26:59 crc kubenswrapper[4967]: I1008 00:26:59.847903 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b742fb05-bba8-4bc5-8ac7-97e1061a8842-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "b742fb05-bba8-4bc5-8ac7-97e1061a8842" (UID: "b742fb05-bba8-4bc5-8ac7-97e1061a8842"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:26:59 crc kubenswrapper[4967]: I1008 00:26:59.849636 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b742fb05-bba8-4bc5-8ac7-97e1061a8842-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "b742fb05-bba8-4bc5-8ac7-97e1061a8842" (UID: "b742fb05-bba8-4bc5-8ac7-97e1061a8842"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:26:59 crc kubenswrapper[4967]: I1008 00:26:59.851534 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b742fb05-bba8-4bc5-8ac7-97e1061a8842-kube-api-access-rwjzq" (OuterVolumeSpecName: "kube-api-access-rwjzq") pod "b742fb05-bba8-4bc5-8ac7-97e1061a8842" (UID: "b742fb05-bba8-4bc5-8ac7-97e1061a8842"). InnerVolumeSpecName "kube-api-access-rwjzq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:26:59 crc kubenswrapper[4967]: I1008 00:26:59.855363 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b742fb05-bba8-4bc5-8ac7-97e1061a8842-builder-dockercfg-r7snr-pull" (OuterVolumeSpecName: "builder-dockercfg-r7snr-pull") pod "b742fb05-bba8-4bc5-8ac7-97e1061a8842" (UID: "b742fb05-bba8-4bc5-8ac7-97e1061a8842"). InnerVolumeSpecName "builder-dockercfg-r7snr-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:26:59 crc kubenswrapper[4967]: I1008 00:26:59.855631 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b742fb05-bba8-4bc5-8ac7-97e1061a8842-builder-dockercfg-r7snr-push" (OuterVolumeSpecName: "builder-dockercfg-r7snr-push") pod "b742fb05-bba8-4bc5-8ac7-97e1061a8842" (UID: "b742fb05-bba8-4bc5-8ac7-97e1061a8842"). InnerVolumeSpecName "builder-dockercfg-r7snr-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:26:59 crc kubenswrapper[4967]: I1008 00:26:59.899572 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b742fb05-bba8-4bc5-8ac7-97e1061a8842-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "b742fb05-bba8-4bc5-8ac7-97e1061a8842" (UID: "b742fb05-bba8-4bc5-8ac7-97e1061a8842"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:26:59 crc kubenswrapper[4967]: I1008 00:26:59.946212 4967 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/b742fb05-bba8-4bc5-8ac7-97e1061a8842-buildworkdir\") on node \"crc\" DevicePath \"\"" Oct 08 00:26:59 crc kubenswrapper[4967]: I1008 00:26:59.946250 4967 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b742fb05-bba8-4bc5-8ac7-97e1061a8842-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 08 00:26:59 crc kubenswrapper[4967]: I1008 00:26:59.946264 4967 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-r7snr-push\" (UniqueName: \"kubernetes.io/secret/b742fb05-bba8-4bc5-8ac7-97e1061a8842-builder-dockercfg-r7snr-push\") on node \"crc\" DevicePath \"\"" Oct 08 00:26:59 crc kubenswrapper[4967]: I1008 00:26:59.946276 4967 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/b742fb05-bba8-4bc5-8ac7-97e1061a8842-buildcachedir\") on node \"crc\" DevicePath \"\"" Oct 08 00:26:59 crc kubenswrapper[4967]: I1008 00:26:59.946287 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rwjzq\" (UniqueName: \"kubernetes.io/projected/b742fb05-bba8-4bc5-8ac7-97e1061a8842-kube-api-access-rwjzq\") on node \"crc\" DevicePath \"\"" Oct 08 00:26:59 crc kubenswrapper[4967]: I1008 00:26:59.946299 4967 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/b742fb05-bba8-4bc5-8ac7-97e1061a8842-build-blob-cache\") on node \"crc\" DevicePath \"\"" Oct 08 00:26:59 crc kubenswrapper[4967]: I1008 00:26:59.946310 4967 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/b742fb05-bba8-4bc5-8ac7-97e1061a8842-container-storage-run\") on node \"crc\" DevicePath \"\"" Oct 08 00:26:59 crc kubenswrapper[4967]: I1008 00:26:59.946321 4967 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/b742fb05-bba8-4bc5-8ac7-97e1061a8842-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Oct 08 00:26:59 crc kubenswrapper[4967]: I1008 00:26:59.946333 4967 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b742fb05-bba8-4bc5-8ac7-97e1061a8842-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 08 00:26:59 crc kubenswrapper[4967]: I1008 00:26:59.946344 4967 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/b742fb05-bba8-4bc5-8ac7-97e1061a8842-build-system-configs\") on node \"crc\" DevicePath \"\"" Oct 08 00:26:59 crc kubenswrapper[4967]: I1008 00:26:59.946383 4967 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-r7snr-pull\" (UniqueName: \"kubernetes.io/secret/b742fb05-bba8-4bc5-8ac7-97e1061a8842-builder-dockercfg-r7snr-pull\") on node \"crc\" DevicePath \"\"" Oct 08 00:27:00 crc kubenswrapper[4967]: I1008 00:27:00.170875 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b742fb05-bba8-4bc5-8ac7-97e1061a8842-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "b742fb05-bba8-4bc5-8ac7-97e1061a8842" (UID: "b742fb05-bba8-4bc5-8ac7-97e1061a8842"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:27:00 crc kubenswrapper[4967]: I1008 00:27:00.250367 4967 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/b742fb05-bba8-4bc5-8ac7-97e1061a8842-container-storage-root\") on node \"crc\" DevicePath \"\"" Oct 08 00:27:00 crc kubenswrapper[4967]: I1008 00:27:00.627468 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_prometheus-webhook-snmp-1-build_b742fb05-bba8-4bc5-8ac7-97e1061a8842/docker-build/0.log" Oct 08 00:27:00 crc kubenswrapper[4967]: I1008 00:27:00.629204 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-1-build" event={"ID":"b742fb05-bba8-4bc5-8ac7-97e1061a8842","Type":"ContainerDied","Data":"a8be1f435a0efc34f9bf4420d630af8251ce95e51b50c711a041676d581159f0"} Oct 08 00:27:00 crc kubenswrapper[4967]: I1008 00:27:00.629282 4967 scope.go:117] "RemoveContainer" containerID="0c89eecbd340ba06c854d218288b3c21bda552c853702d29d706dd38c00e64d2" Oct 08 00:27:00 crc kubenswrapper[4967]: I1008 00:27:00.629511 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-webhook-snmp-1-build" Oct 08 00:27:00 crc kubenswrapper[4967]: I1008 00:27:00.663245 4967 scope.go:117] "RemoveContainer" containerID="8f97f93a2bf10211cadfd3b6fc1efd07aa96898fb604a1d7c39e29f803acfb2d" Oct 08 00:27:00 crc kubenswrapper[4967]: I1008 00:27:00.687828 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/prometheus-webhook-snmp-1-build"] Oct 08 00:27:00 crc kubenswrapper[4967]: I1008 00:27:00.695870 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["service-telemetry/prometheus-webhook-snmp-1-build"] Oct 08 00:27:00 crc kubenswrapper[4967]: I1008 00:27:00.968212 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/prometheus-webhook-snmp-2-build"] Oct 08 00:27:00 crc kubenswrapper[4967]: E1008 00:27:00.968576 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b742fb05-bba8-4bc5-8ac7-97e1061a8842" containerName="docker-build" Oct 08 00:27:00 crc kubenswrapper[4967]: I1008 00:27:00.968648 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="b742fb05-bba8-4bc5-8ac7-97e1061a8842" containerName="docker-build" Oct 08 00:27:00 crc kubenswrapper[4967]: E1008 00:27:00.968733 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b742fb05-bba8-4bc5-8ac7-97e1061a8842" containerName="manage-dockerfile" Oct 08 00:27:00 crc kubenswrapper[4967]: I1008 00:27:00.968748 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="b742fb05-bba8-4bc5-8ac7-97e1061a8842" containerName="manage-dockerfile" Oct 08 00:27:00 crc kubenswrapper[4967]: I1008 00:27:00.968997 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="b742fb05-bba8-4bc5-8ac7-97e1061a8842" containerName="docker-build" Oct 08 00:27:00 crc kubenswrapper[4967]: I1008 00:27:00.970930 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-webhook-snmp-2-build" Oct 08 00:27:00 crc kubenswrapper[4967]: I1008 00:27:00.974451 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"prometheus-webhook-snmp-2-sys-config" Oct 08 00:27:00 crc kubenswrapper[4967]: I1008 00:27:00.976297 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"prometheus-webhook-snmp-2-ca" Oct 08 00:27:00 crc kubenswrapper[4967]: I1008 00:27:00.976392 4967 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-r7snr" Oct 08 00:27:00 crc kubenswrapper[4967]: I1008 00:27:00.976394 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"prometheus-webhook-snmp-2-global-ca" Oct 08 00:27:00 crc kubenswrapper[4967]: I1008 00:27:00.997971 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/prometheus-webhook-snmp-2-build"] Oct 08 00:27:01 crc kubenswrapper[4967]: I1008 00:27:01.063742 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/dcfdb561-abc8-465e-9494-23950be122ce-container-storage-root\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"dcfdb561-abc8-465e-9494-23950be122ce\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Oct 08 00:27:01 crc kubenswrapper[4967]: I1008 00:27:01.064121 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/dcfdb561-abc8-465e-9494-23950be122ce-build-system-configs\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"dcfdb561-abc8-465e-9494-23950be122ce\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Oct 08 00:27:01 crc kubenswrapper[4967]: I1008 00:27:01.064346 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/dcfdb561-abc8-465e-9494-23950be122ce-container-storage-run\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"dcfdb561-abc8-465e-9494-23950be122ce\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Oct 08 00:27:01 crc kubenswrapper[4967]: I1008 00:27:01.064583 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/dcfdb561-abc8-465e-9494-23950be122ce-buildworkdir\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"dcfdb561-abc8-465e-9494-23950be122ce\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Oct 08 00:27:01 crc kubenswrapper[4967]: I1008 00:27:01.064852 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/dcfdb561-abc8-465e-9494-23950be122ce-build-proxy-ca-bundles\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"dcfdb561-abc8-465e-9494-23950be122ce\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Oct 08 00:27:01 crc kubenswrapper[4967]: I1008 00:27:01.065037 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/dcfdb561-abc8-465e-9494-23950be122ce-buildcachedir\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"dcfdb561-abc8-465e-9494-23950be122ce\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Oct 08 00:27:01 crc kubenswrapper[4967]: I1008 00:27:01.065198 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gftkc\" (UniqueName: \"kubernetes.io/projected/dcfdb561-abc8-465e-9494-23950be122ce-kube-api-access-gftkc\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"dcfdb561-abc8-465e-9494-23950be122ce\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Oct 08 00:27:01 crc kubenswrapper[4967]: I1008 00:27:01.065360 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-r7snr-push\" (UniqueName: \"kubernetes.io/secret/dcfdb561-abc8-465e-9494-23950be122ce-builder-dockercfg-r7snr-push\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"dcfdb561-abc8-465e-9494-23950be122ce\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Oct 08 00:27:01 crc kubenswrapper[4967]: I1008 00:27:01.065535 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/dcfdb561-abc8-465e-9494-23950be122ce-build-blob-cache\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"dcfdb561-abc8-465e-9494-23950be122ce\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Oct 08 00:27:01 crc kubenswrapper[4967]: I1008 00:27:01.065775 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/dcfdb561-abc8-465e-9494-23950be122ce-build-ca-bundles\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"dcfdb561-abc8-465e-9494-23950be122ce\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Oct 08 00:27:01 crc kubenswrapper[4967]: I1008 00:27:01.065968 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/dcfdb561-abc8-465e-9494-23950be122ce-node-pullsecrets\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"dcfdb561-abc8-465e-9494-23950be122ce\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Oct 08 00:27:01 crc kubenswrapper[4967]: I1008 00:27:01.066129 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-r7snr-pull\" (UniqueName: \"kubernetes.io/secret/dcfdb561-abc8-465e-9494-23950be122ce-builder-dockercfg-r7snr-pull\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"dcfdb561-abc8-465e-9494-23950be122ce\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Oct 08 00:27:01 crc kubenswrapper[4967]: I1008 00:27:01.167198 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/dcfdb561-abc8-465e-9494-23950be122ce-buildworkdir\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"dcfdb561-abc8-465e-9494-23950be122ce\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Oct 08 00:27:01 crc kubenswrapper[4967]: I1008 00:27:01.167246 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/dcfdb561-abc8-465e-9494-23950be122ce-build-proxy-ca-bundles\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"dcfdb561-abc8-465e-9494-23950be122ce\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Oct 08 00:27:01 crc kubenswrapper[4967]: I1008 00:27:01.167275 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/dcfdb561-abc8-465e-9494-23950be122ce-buildcachedir\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"dcfdb561-abc8-465e-9494-23950be122ce\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Oct 08 00:27:01 crc kubenswrapper[4967]: I1008 00:27:01.167306 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gftkc\" (UniqueName: \"kubernetes.io/projected/dcfdb561-abc8-465e-9494-23950be122ce-kube-api-access-gftkc\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"dcfdb561-abc8-465e-9494-23950be122ce\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Oct 08 00:27:01 crc kubenswrapper[4967]: I1008 00:27:01.167335 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-r7snr-push\" (UniqueName: \"kubernetes.io/secret/dcfdb561-abc8-465e-9494-23950be122ce-builder-dockercfg-r7snr-push\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"dcfdb561-abc8-465e-9494-23950be122ce\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Oct 08 00:27:01 crc kubenswrapper[4967]: I1008 00:27:01.167388 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/dcfdb561-abc8-465e-9494-23950be122ce-build-blob-cache\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"dcfdb561-abc8-465e-9494-23950be122ce\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Oct 08 00:27:01 crc kubenswrapper[4967]: I1008 00:27:01.167444 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/dcfdb561-abc8-465e-9494-23950be122ce-build-ca-bundles\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"dcfdb561-abc8-465e-9494-23950be122ce\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Oct 08 00:27:01 crc kubenswrapper[4967]: I1008 00:27:01.167499 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/dcfdb561-abc8-465e-9494-23950be122ce-node-pullsecrets\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"dcfdb561-abc8-465e-9494-23950be122ce\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Oct 08 00:27:01 crc kubenswrapper[4967]: I1008 00:27:01.167544 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-r7snr-pull\" (UniqueName: \"kubernetes.io/secret/dcfdb561-abc8-465e-9494-23950be122ce-builder-dockercfg-r7snr-pull\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"dcfdb561-abc8-465e-9494-23950be122ce\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Oct 08 00:27:01 crc kubenswrapper[4967]: I1008 00:27:01.167543 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/dcfdb561-abc8-465e-9494-23950be122ce-buildcachedir\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"dcfdb561-abc8-465e-9494-23950be122ce\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Oct 08 00:27:01 crc kubenswrapper[4967]: I1008 00:27:01.167572 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/dcfdb561-abc8-465e-9494-23950be122ce-container-storage-root\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"dcfdb561-abc8-465e-9494-23950be122ce\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Oct 08 00:27:01 crc kubenswrapper[4967]: I1008 00:27:01.167805 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/dcfdb561-abc8-465e-9494-23950be122ce-buildworkdir\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"dcfdb561-abc8-465e-9494-23950be122ce\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Oct 08 00:27:01 crc kubenswrapper[4967]: I1008 00:27:01.167818 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/dcfdb561-abc8-465e-9494-23950be122ce-build-system-configs\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"dcfdb561-abc8-465e-9494-23950be122ce\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Oct 08 00:27:01 crc kubenswrapper[4967]: I1008 00:27:01.167930 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/dcfdb561-abc8-465e-9494-23950be122ce-container-storage-run\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"dcfdb561-abc8-465e-9494-23950be122ce\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Oct 08 00:27:01 crc kubenswrapper[4967]: I1008 00:27:01.167982 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/dcfdb561-abc8-465e-9494-23950be122ce-container-storage-root\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"dcfdb561-abc8-465e-9494-23950be122ce\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Oct 08 00:27:01 crc kubenswrapper[4967]: I1008 00:27:01.168050 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/dcfdb561-abc8-465e-9494-23950be122ce-node-pullsecrets\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"dcfdb561-abc8-465e-9494-23950be122ce\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Oct 08 00:27:01 crc kubenswrapper[4967]: I1008 00:27:01.168919 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/dcfdb561-abc8-465e-9494-23950be122ce-build-proxy-ca-bundles\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"dcfdb561-abc8-465e-9494-23950be122ce\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Oct 08 00:27:01 crc kubenswrapper[4967]: I1008 00:27:01.169841 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/dcfdb561-abc8-465e-9494-23950be122ce-container-storage-run\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"dcfdb561-abc8-465e-9494-23950be122ce\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Oct 08 00:27:01 crc kubenswrapper[4967]: I1008 00:27:01.170012 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/dcfdb561-abc8-465e-9494-23950be122ce-build-blob-cache\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"dcfdb561-abc8-465e-9494-23950be122ce\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Oct 08 00:27:01 crc kubenswrapper[4967]: I1008 00:27:01.170237 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/dcfdb561-abc8-465e-9494-23950be122ce-build-ca-bundles\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"dcfdb561-abc8-465e-9494-23950be122ce\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Oct 08 00:27:01 crc kubenswrapper[4967]: I1008 00:27:01.170244 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/dcfdb561-abc8-465e-9494-23950be122ce-build-system-configs\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"dcfdb561-abc8-465e-9494-23950be122ce\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Oct 08 00:27:01 crc kubenswrapper[4967]: I1008 00:27:01.180156 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-r7snr-pull\" (UniqueName: \"kubernetes.io/secret/dcfdb561-abc8-465e-9494-23950be122ce-builder-dockercfg-r7snr-pull\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"dcfdb561-abc8-465e-9494-23950be122ce\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Oct 08 00:27:01 crc kubenswrapper[4967]: I1008 00:27:01.181427 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-r7snr-push\" (UniqueName: \"kubernetes.io/secret/dcfdb561-abc8-465e-9494-23950be122ce-builder-dockercfg-r7snr-push\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"dcfdb561-abc8-465e-9494-23950be122ce\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Oct 08 00:27:01 crc kubenswrapper[4967]: I1008 00:27:01.191006 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gftkc\" (UniqueName: \"kubernetes.io/projected/dcfdb561-abc8-465e-9494-23950be122ce-kube-api-access-gftkc\") pod \"prometheus-webhook-snmp-2-build\" (UID: \"dcfdb561-abc8-465e-9494-23950be122ce\") " pod="service-telemetry/prometheus-webhook-snmp-2-build" Oct 08 00:27:01 crc kubenswrapper[4967]: I1008 00:27:01.303425 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-webhook-snmp-2-build" Oct 08 00:27:01 crc kubenswrapper[4967]: I1008 00:27:01.579040 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/prometheus-webhook-snmp-2-build"] Oct 08 00:27:01 crc kubenswrapper[4967]: I1008 00:27:01.636901 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-2-build" event={"ID":"dcfdb561-abc8-465e-9494-23950be122ce","Type":"ContainerStarted","Data":"9a656cb1d4e1dbb16bf3ce6f472357b95fdf010610b21080daa2aec0383310fa"} Oct 08 00:27:02 crc kubenswrapper[4967]: I1008 00:27:02.129470 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b742fb05-bba8-4bc5-8ac7-97e1061a8842" path="/var/lib/kubelet/pods/b742fb05-bba8-4bc5-8ac7-97e1061a8842/volumes" Oct 08 00:27:02 crc kubenswrapper[4967]: I1008 00:27:02.653996 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-2-build" event={"ID":"dcfdb561-abc8-465e-9494-23950be122ce","Type":"ContainerStarted","Data":"58a0d96e0f88987c5dd57417f4cfec1c6b4d92ade5a023ff0cfb8cc0f3dd0539"} Oct 08 00:27:03 crc kubenswrapper[4967]: I1008 00:27:03.664515 4967 generic.go:334] "Generic (PLEG): container finished" podID="dcfdb561-abc8-465e-9494-23950be122ce" containerID="58a0d96e0f88987c5dd57417f4cfec1c6b4d92ade5a023ff0cfb8cc0f3dd0539" exitCode=0 Oct 08 00:27:03 crc kubenswrapper[4967]: I1008 00:27:03.664616 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-2-build" event={"ID":"dcfdb561-abc8-465e-9494-23950be122ce","Type":"ContainerDied","Data":"58a0d96e0f88987c5dd57417f4cfec1c6b4d92ade5a023ff0cfb8cc0f3dd0539"} Oct 08 00:27:04 crc kubenswrapper[4967]: I1008 00:27:04.675338 4967 generic.go:334] "Generic (PLEG): container finished" podID="dcfdb561-abc8-465e-9494-23950be122ce" containerID="7013f59c4c259220801c3a1f43aa2d73fa7c06891c30da4b471beb05c23ab07e" exitCode=0 Oct 08 00:27:04 crc kubenswrapper[4967]: I1008 00:27:04.675396 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-2-build" event={"ID":"dcfdb561-abc8-465e-9494-23950be122ce","Type":"ContainerDied","Data":"7013f59c4c259220801c3a1f43aa2d73fa7c06891c30da4b471beb05c23ab07e"} Oct 08 00:27:04 crc kubenswrapper[4967]: I1008 00:27:04.722014 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_prometheus-webhook-snmp-2-build_dcfdb561-abc8-465e-9494-23950be122ce/manage-dockerfile/0.log" Oct 08 00:27:05 crc kubenswrapper[4967]: I1008 00:27:05.690949 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-2-build" event={"ID":"dcfdb561-abc8-465e-9494-23950be122ce","Type":"ContainerStarted","Data":"10072f49c71434e9eb6329e0773db603293403e0028b47b8e531e06d6ae5e044"} Oct 08 00:27:05 crc kubenswrapper[4967]: I1008 00:27:05.727370 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/prometheus-webhook-snmp-2-build" podStartSLOduration=5.727332449 podStartE2EDuration="5.727332449s" podCreationTimestamp="2025-10-08 00:27:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:27:05.72406248 +0000 UTC m=+1224.374597219" watchObservedRunningTime="2025-10-08 00:27:05.727332449 +0000 UTC m=+1224.377867128" Oct 08 00:27:06 crc kubenswrapper[4967]: I1008 00:27:06.920431 4967 patch_prober.go:28] interesting pod/machine-config-daemon-xvdtp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 00:27:06 crc kubenswrapper[4967]: I1008 00:27:06.920512 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" podUID="9461162a-c27d-465e-9ca7-2d3936877cd8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 00:27:36 crc kubenswrapper[4967]: I1008 00:27:36.920497 4967 patch_prober.go:28] interesting pod/machine-config-daemon-xvdtp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 00:27:36 crc kubenswrapper[4967]: I1008 00:27:36.921243 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" podUID="9461162a-c27d-465e-9ca7-2d3936877cd8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 00:27:36 crc kubenswrapper[4967]: I1008 00:27:36.921313 4967 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" Oct 08 00:27:36 crc kubenswrapper[4967]: I1008 00:27:36.922229 4967 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a9bcd32e3405379dbe88886537687c4d0982cbf18aecd6bc1e1ae62026f21458"} pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 00:27:36 crc kubenswrapper[4967]: I1008 00:27:36.922331 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" podUID="9461162a-c27d-465e-9ca7-2d3936877cd8" containerName="machine-config-daemon" containerID="cri-o://a9bcd32e3405379dbe88886537687c4d0982cbf18aecd6bc1e1ae62026f21458" gracePeriod=600 Oct 08 00:27:37 crc kubenswrapper[4967]: I1008 00:27:37.957238 4967 generic.go:334] "Generic (PLEG): container finished" podID="9461162a-c27d-465e-9ca7-2d3936877cd8" containerID="a9bcd32e3405379dbe88886537687c4d0982cbf18aecd6bc1e1ae62026f21458" exitCode=0 Oct 08 00:27:37 crc kubenswrapper[4967]: I1008 00:27:37.957327 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" event={"ID":"9461162a-c27d-465e-9ca7-2d3936877cd8","Type":"ContainerDied","Data":"a9bcd32e3405379dbe88886537687c4d0982cbf18aecd6bc1e1ae62026f21458"} Oct 08 00:27:37 crc kubenswrapper[4967]: I1008 00:27:37.957920 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" event={"ID":"9461162a-c27d-465e-9ca7-2d3936877cd8","Type":"ContainerStarted","Data":"5e1e0751dcae2b0e4bad8c7f2220f3bca21a69c9cccc3dd960fa9770bb1adf17"} Oct 08 00:27:37 crc kubenswrapper[4967]: I1008 00:27:37.957941 4967 scope.go:117] "RemoveContainer" containerID="bd8678b8254640836f0db760c5c6a0da933718c28fc5033465f9875f7bd725b1" Oct 08 00:27:50 crc kubenswrapper[4967]: E1008 00:27:50.110856 4967 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddcfdb561_abc8_465e_9494_23950be122ce.slice/buildah-buildah317276172\": RecentStats: unable to find data in memory cache]" Oct 08 00:28:03 crc kubenswrapper[4967]: I1008 00:28:03.215868 4967 generic.go:334] "Generic (PLEG): container finished" podID="dcfdb561-abc8-465e-9494-23950be122ce" containerID="10072f49c71434e9eb6329e0773db603293403e0028b47b8e531e06d6ae5e044" exitCode=0 Oct 08 00:28:03 crc kubenswrapper[4967]: I1008 00:28:03.215989 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-2-build" event={"ID":"dcfdb561-abc8-465e-9494-23950be122ce","Type":"ContainerDied","Data":"10072f49c71434e9eb6329e0773db603293403e0028b47b8e531e06d6ae5e044"} Oct 08 00:28:04 crc kubenswrapper[4967]: I1008 00:28:04.471096 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-webhook-snmp-2-build" Oct 08 00:28:04 crc kubenswrapper[4967]: I1008 00:28:04.578751 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/dcfdb561-abc8-465e-9494-23950be122ce-container-storage-root\") pod \"dcfdb561-abc8-465e-9494-23950be122ce\" (UID: \"dcfdb561-abc8-465e-9494-23950be122ce\") " Oct 08 00:28:04 crc kubenswrapper[4967]: I1008 00:28:04.578794 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/dcfdb561-abc8-465e-9494-23950be122ce-build-ca-bundles\") pod \"dcfdb561-abc8-465e-9494-23950be122ce\" (UID: \"dcfdb561-abc8-465e-9494-23950be122ce\") " Oct 08 00:28:04 crc kubenswrapper[4967]: I1008 00:28:04.578864 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/dcfdb561-abc8-465e-9494-23950be122ce-buildcachedir\") pod \"dcfdb561-abc8-465e-9494-23950be122ce\" (UID: \"dcfdb561-abc8-465e-9494-23950be122ce\") " Oct 08 00:28:04 crc kubenswrapper[4967]: I1008 00:28:04.578912 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gftkc\" (UniqueName: \"kubernetes.io/projected/dcfdb561-abc8-465e-9494-23950be122ce-kube-api-access-gftkc\") pod \"dcfdb561-abc8-465e-9494-23950be122ce\" (UID: \"dcfdb561-abc8-465e-9494-23950be122ce\") " Oct 08 00:28:04 crc kubenswrapper[4967]: I1008 00:28:04.578931 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-r7snr-push\" (UniqueName: \"kubernetes.io/secret/dcfdb561-abc8-465e-9494-23950be122ce-builder-dockercfg-r7snr-push\") pod \"dcfdb561-abc8-465e-9494-23950be122ce\" (UID: \"dcfdb561-abc8-465e-9494-23950be122ce\") " Oct 08 00:28:04 crc kubenswrapper[4967]: I1008 00:28:04.578952 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/dcfdb561-abc8-465e-9494-23950be122ce-build-system-configs\") pod \"dcfdb561-abc8-465e-9494-23950be122ce\" (UID: \"dcfdb561-abc8-465e-9494-23950be122ce\") " Oct 08 00:28:04 crc kubenswrapper[4967]: I1008 00:28:04.578975 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/dcfdb561-abc8-465e-9494-23950be122ce-container-storage-run\") pod \"dcfdb561-abc8-465e-9494-23950be122ce\" (UID: \"dcfdb561-abc8-465e-9494-23950be122ce\") " Oct 08 00:28:04 crc kubenswrapper[4967]: I1008 00:28:04.578973 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/dcfdb561-abc8-465e-9494-23950be122ce-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "dcfdb561-abc8-465e-9494-23950be122ce" (UID: "dcfdb561-abc8-465e-9494-23950be122ce"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 00:28:04 crc kubenswrapper[4967]: I1008 00:28:04.579030 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/dcfdb561-abc8-465e-9494-23950be122ce-build-blob-cache\") pod \"dcfdb561-abc8-465e-9494-23950be122ce\" (UID: \"dcfdb561-abc8-465e-9494-23950be122ce\") " Oct 08 00:28:04 crc kubenswrapper[4967]: I1008 00:28:04.579049 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/dcfdb561-abc8-465e-9494-23950be122ce-buildworkdir\") pod \"dcfdb561-abc8-465e-9494-23950be122ce\" (UID: \"dcfdb561-abc8-465e-9494-23950be122ce\") " Oct 08 00:28:04 crc kubenswrapper[4967]: I1008 00:28:04.579081 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/dcfdb561-abc8-465e-9494-23950be122ce-build-proxy-ca-bundles\") pod \"dcfdb561-abc8-465e-9494-23950be122ce\" (UID: \"dcfdb561-abc8-465e-9494-23950be122ce\") " Oct 08 00:28:04 crc kubenswrapper[4967]: I1008 00:28:04.579096 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/dcfdb561-abc8-465e-9494-23950be122ce-node-pullsecrets\") pod \"dcfdb561-abc8-465e-9494-23950be122ce\" (UID: \"dcfdb561-abc8-465e-9494-23950be122ce\") " Oct 08 00:28:04 crc kubenswrapper[4967]: I1008 00:28:04.579127 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-r7snr-pull\" (UniqueName: \"kubernetes.io/secret/dcfdb561-abc8-465e-9494-23950be122ce-builder-dockercfg-r7snr-pull\") pod \"dcfdb561-abc8-465e-9494-23950be122ce\" (UID: \"dcfdb561-abc8-465e-9494-23950be122ce\") " Oct 08 00:28:04 crc kubenswrapper[4967]: I1008 00:28:04.579369 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/dcfdb561-abc8-465e-9494-23950be122ce-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "dcfdb561-abc8-465e-9494-23950be122ce" (UID: "dcfdb561-abc8-465e-9494-23950be122ce"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 00:28:04 crc kubenswrapper[4967]: I1008 00:28:04.579410 4967 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/dcfdb561-abc8-465e-9494-23950be122ce-buildcachedir\") on node \"crc\" DevicePath \"\"" Oct 08 00:28:04 crc kubenswrapper[4967]: I1008 00:28:04.579558 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dcfdb561-abc8-465e-9494-23950be122ce-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "dcfdb561-abc8-465e-9494-23950be122ce" (UID: "dcfdb561-abc8-465e-9494-23950be122ce"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:28:04 crc kubenswrapper[4967]: I1008 00:28:04.580222 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dcfdb561-abc8-465e-9494-23950be122ce-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "dcfdb561-abc8-465e-9494-23950be122ce" (UID: "dcfdb561-abc8-465e-9494-23950be122ce"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:28:04 crc kubenswrapper[4967]: I1008 00:28:04.580325 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dcfdb561-abc8-465e-9494-23950be122ce-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "dcfdb561-abc8-465e-9494-23950be122ce" (UID: "dcfdb561-abc8-465e-9494-23950be122ce"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:28:04 crc kubenswrapper[4967]: I1008 00:28:04.582593 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dcfdb561-abc8-465e-9494-23950be122ce-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "dcfdb561-abc8-465e-9494-23950be122ce" (UID: "dcfdb561-abc8-465e-9494-23950be122ce"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:28:04 crc kubenswrapper[4967]: I1008 00:28:04.583979 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dcfdb561-abc8-465e-9494-23950be122ce-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "dcfdb561-abc8-465e-9494-23950be122ce" (UID: "dcfdb561-abc8-465e-9494-23950be122ce"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:28:04 crc kubenswrapper[4967]: I1008 00:28:04.584476 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dcfdb561-abc8-465e-9494-23950be122ce-kube-api-access-gftkc" (OuterVolumeSpecName: "kube-api-access-gftkc") pod "dcfdb561-abc8-465e-9494-23950be122ce" (UID: "dcfdb561-abc8-465e-9494-23950be122ce"). InnerVolumeSpecName "kube-api-access-gftkc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:28:04 crc kubenswrapper[4967]: I1008 00:28:04.584845 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dcfdb561-abc8-465e-9494-23950be122ce-builder-dockercfg-r7snr-push" (OuterVolumeSpecName: "builder-dockercfg-r7snr-push") pod "dcfdb561-abc8-465e-9494-23950be122ce" (UID: "dcfdb561-abc8-465e-9494-23950be122ce"). InnerVolumeSpecName "builder-dockercfg-r7snr-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:28:04 crc kubenswrapper[4967]: I1008 00:28:04.585490 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dcfdb561-abc8-465e-9494-23950be122ce-builder-dockercfg-r7snr-pull" (OuterVolumeSpecName: "builder-dockercfg-r7snr-pull") pod "dcfdb561-abc8-465e-9494-23950be122ce" (UID: "dcfdb561-abc8-465e-9494-23950be122ce"). InnerVolumeSpecName "builder-dockercfg-r7snr-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:28:04 crc kubenswrapper[4967]: I1008 00:28:04.681077 4967 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/dcfdb561-abc8-465e-9494-23950be122ce-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 08 00:28:04 crc kubenswrapper[4967]: I1008 00:28:04.681109 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gftkc\" (UniqueName: \"kubernetes.io/projected/dcfdb561-abc8-465e-9494-23950be122ce-kube-api-access-gftkc\") on node \"crc\" DevicePath \"\"" Oct 08 00:28:04 crc kubenswrapper[4967]: I1008 00:28:04.681119 4967 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-r7snr-push\" (UniqueName: \"kubernetes.io/secret/dcfdb561-abc8-465e-9494-23950be122ce-builder-dockercfg-r7snr-push\") on node \"crc\" DevicePath \"\"" Oct 08 00:28:04 crc kubenswrapper[4967]: I1008 00:28:04.681129 4967 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/dcfdb561-abc8-465e-9494-23950be122ce-build-system-configs\") on node \"crc\" DevicePath \"\"" Oct 08 00:28:04 crc kubenswrapper[4967]: I1008 00:28:04.681137 4967 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/dcfdb561-abc8-465e-9494-23950be122ce-container-storage-run\") on node \"crc\" DevicePath \"\"" Oct 08 00:28:04 crc kubenswrapper[4967]: I1008 00:28:04.681146 4967 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/dcfdb561-abc8-465e-9494-23950be122ce-buildworkdir\") on node \"crc\" DevicePath \"\"" Oct 08 00:28:04 crc kubenswrapper[4967]: I1008 00:28:04.681153 4967 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/dcfdb561-abc8-465e-9494-23950be122ce-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 08 00:28:04 crc kubenswrapper[4967]: I1008 00:28:04.681162 4967 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/dcfdb561-abc8-465e-9494-23950be122ce-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Oct 08 00:28:04 crc kubenswrapper[4967]: I1008 00:28:04.681169 4967 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-r7snr-pull\" (UniqueName: \"kubernetes.io/secret/dcfdb561-abc8-465e-9494-23950be122ce-builder-dockercfg-r7snr-pull\") on node \"crc\" DevicePath \"\"" Oct 08 00:28:04 crc kubenswrapper[4967]: I1008 00:28:04.687476 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dcfdb561-abc8-465e-9494-23950be122ce-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "dcfdb561-abc8-465e-9494-23950be122ce" (UID: "dcfdb561-abc8-465e-9494-23950be122ce"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:28:04 crc kubenswrapper[4967]: I1008 00:28:04.781968 4967 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/dcfdb561-abc8-465e-9494-23950be122ce-build-blob-cache\") on node \"crc\" DevicePath \"\"" Oct 08 00:28:05 crc kubenswrapper[4967]: I1008 00:28:05.237330 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-webhook-snmp-2-build" event={"ID":"dcfdb561-abc8-465e-9494-23950be122ce","Type":"ContainerDied","Data":"9a656cb1d4e1dbb16bf3ce6f472357b95fdf010610b21080daa2aec0383310fa"} Oct 08 00:28:05 crc kubenswrapper[4967]: I1008 00:28:05.237406 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9a656cb1d4e1dbb16bf3ce6f472357b95fdf010610b21080daa2aec0383310fa" Oct 08 00:28:05 crc kubenswrapper[4967]: I1008 00:28:05.238031 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-webhook-snmp-2-build" Oct 08 00:28:05 crc kubenswrapper[4967]: I1008 00:28:05.598903 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dcfdb561-abc8-465e-9494-23950be122ce-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "dcfdb561-abc8-465e-9494-23950be122ce" (UID: "dcfdb561-abc8-465e-9494-23950be122ce"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:28:05 crc kubenswrapper[4967]: I1008 00:28:05.602040 4967 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/dcfdb561-abc8-465e-9494-23950be122ce-container-storage-root\") on node \"crc\" DevicePath \"\"" Oct 08 00:28:10 crc kubenswrapper[4967]: I1008 00:28:10.355709 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/smart-gateway-operator-7d98fcf998-q8f6v"] Oct 08 00:28:10 crc kubenswrapper[4967]: E1008 00:28:10.356710 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dcfdb561-abc8-465e-9494-23950be122ce" containerName="git-clone" Oct 08 00:28:10 crc kubenswrapper[4967]: I1008 00:28:10.356733 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="dcfdb561-abc8-465e-9494-23950be122ce" containerName="git-clone" Oct 08 00:28:10 crc kubenswrapper[4967]: E1008 00:28:10.356760 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dcfdb561-abc8-465e-9494-23950be122ce" containerName="docker-build" Oct 08 00:28:10 crc kubenswrapper[4967]: I1008 00:28:10.356772 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="dcfdb561-abc8-465e-9494-23950be122ce" containerName="docker-build" Oct 08 00:28:10 crc kubenswrapper[4967]: E1008 00:28:10.356803 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dcfdb561-abc8-465e-9494-23950be122ce" containerName="manage-dockerfile" Oct 08 00:28:10 crc kubenswrapper[4967]: I1008 00:28:10.356814 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="dcfdb561-abc8-465e-9494-23950be122ce" containerName="manage-dockerfile" Oct 08 00:28:10 crc kubenswrapper[4967]: I1008 00:28:10.357015 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="dcfdb561-abc8-465e-9494-23950be122ce" containerName="docker-build" Oct 08 00:28:10 crc kubenswrapper[4967]: I1008 00:28:10.357719 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-7d98fcf998-q8f6v" Oct 08 00:28:10 crc kubenswrapper[4967]: I1008 00:28:10.359816 4967 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"smart-gateway-operator-dockercfg-npvfs" Oct 08 00:28:10 crc kubenswrapper[4967]: I1008 00:28:10.362489 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/smart-gateway-operator-7d98fcf998-q8f6v"] Oct 08 00:28:10 crc kubenswrapper[4967]: I1008 00:28:10.468954 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"runner\" (UniqueName: \"kubernetes.io/empty-dir/5bc8d2b2-4afa-44de-a704-abacdd934b28-runner\") pod \"smart-gateway-operator-7d98fcf998-q8f6v\" (UID: \"5bc8d2b2-4afa-44de-a704-abacdd934b28\") " pod="service-telemetry/smart-gateway-operator-7d98fcf998-q8f6v" Oct 08 00:28:10 crc kubenswrapper[4967]: I1008 00:28:10.469012 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c4n4r\" (UniqueName: \"kubernetes.io/projected/5bc8d2b2-4afa-44de-a704-abacdd934b28-kube-api-access-c4n4r\") pod \"smart-gateway-operator-7d98fcf998-q8f6v\" (UID: \"5bc8d2b2-4afa-44de-a704-abacdd934b28\") " pod="service-telemetry/smart-gateway-operator-7d98fcf998-q8f6v" Oct 08 00:28:10 crc kubenswrapper[4967]: I1008 00:28:10.569868 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"runner\" (UniqueName: \"kubernetes.io/empty-dir/5bc8d2b2-4afa-44de-a704-abacdd934b28-runner\") pod \"smart-gateway-operator-7d98fcf998-q8f6v\" (UID: \"5bc8d2b2-4afa-44de-a704-abacdd934b28\") " pod="service-telemetry/smart-gateway-operator-7d98fcf998-q8f6v" Oct 08 00:28:10 crc kubenswrapper[4967]: I1008 00:28:10.569939 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c4n4r\" (UniqueName: \"kubernetes.io/projected/5bc8d2b2-4afa-44de-a704-abacdd934b28-kube-api-access-c4n4r\") pod \"smart-gateway-operator-7d98fcf998-q8f6v\" (UID: \"5bc8d2b2-4afa-44de-a704-abacdd934b28\") " pod="service-telemetry/smart-gateway-operator-7d98fcf998-q8f6v" Oct 08 00:28:10 crc kubenswrapper[4967]: I1008 00:28:10.570299 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"runner\" (UniqueName: \"kubernetes.io/empty-dir/5bc8d2b2-4afa-44de-a704-abacdd934b28-runner\") pod \"smart-gateway-operator-7d98fcf998-q8f6v\" (UID: \"5bc8d2b2-4afa-44de-a704-abacdd934b28\") " pod="service-telemetry/smart-gateway-operator-7d98fcf998-q8f6v" Oct 08 00:28:10 crc kubenswrapper[4967]: I1008 00:28:10.594510 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c4n4r\" (UniqueName: \"kubernetes.io/projected/5bc8d2b2-4afa-44de-a704-abacdd934b28-kube-api-access-c4n4r\") pod \"smart-gateway-operator-7d98fcf998-q8f6v\" (UID: \"5bc8d2b2-4afa-44de-a704-abacdd934b28\") " pod="service-telemetry/smart-gateway-operator-7d98fcf998-q8f6v" Oct 08 00:28:10 crc kubenswrapper[4967]: I1008 00:28:10.676373 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-7d98fcf998-q8f6v" Oct 08 00:28:10 crc kubenswrapper[4967]: I1008 00:28:10.922549 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/smart-gateway-operator-7d98fcf998-q8f6v"] Oct 08 00:28:10 crc kubenswrapper[4967]: I1008 00:28:10.931400 4967 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 00:28:11 crc kubenswrapper[4967]: I1008 00:28:11.290166 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-7d98fcf998-q8f6v" event={"ID":"5bc8d2b2-4afa-44de-a704-abacdd934b28","Type":"ContainerStarted","Data":"66b2e71a229cf7a71053cfa3b8dc89a2897e8a267c8b99b80b8a82762deeb80f"} Oct 08 00:28:16 crc kubenswrapper[4967]: I1008 00:28:16.532898 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/service-telemetry-operator-56dc879589-jp9r2"] Oct 08 00:28:16 crc kubenswrapper[4967]: I1008 00:28:16.534648 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-56dc879589-jp9r2" Oct 08 00:28:16 crc kubenswrapper[4967]: I1008 00:28:16.537210 4967 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"service-telemetry-operator-dockercfg-fbmns" Oct 08 00:28:16 crc kubenswrapper[4967]: I1008 00:28:16.550532 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sk96r\" (UniqueName: \"kubernetes.io/projected/5f3f5df9-b819-464d-b7ca-2dea8ceab64d-kube-api-access-sk96r\") pod \"service-telemetry-operator-56dc879589-jp9r2\" (UID: \"5f3f5df9-b819-464d-b7ca-2dea8ceab64d\") " pod="service-telemetry/service-telemetry-operator-56dc879589-jp9r2" Oct 08 00:28:16 crc kubenswrapper[4967]: I1008 00:28:16.550593 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"runner\" (UniqueName: \"kubernetes.io/empty-dir/5f3f5df9-b819-464d-b7ca-2dea8ceab64d-runner\") pod \"service-telemetry-operator-56dc879589-jp9r2\" (UID: \"5f3f5df9-b819-464d-b7ca-2dea8ceab64d\") " pod="service-telemetry/service-telemetry-operator-56dc879589-jp9r2" Oct 08 00:28:16 crc kubenswrapper[4967]: I1008 00:28:16.559026 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-56dc879589-jp9r2"] Oct 08 00:28:16 crc kubenswrapper[4967]: I1008 00:28:16.652459 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sk96r\" (UniqueName: \"kubernetes.io/projected/5f3f5df9-b819-464d-b7ca-2dea8ceab64d-kube-api-access-sk96r\") pod \"service-telemetry-operator-56dc879589-jp9r2\" (UID: \"5f3f5df9-b819-464d-b7ca-2dea8ceab64d\") " pod="service-telemetry/service-telemetry-operator-56dc879589-jp9r2" Oct 08 00:28:16 crc kubenswrapper[4967]: I1008 00:28:16.652516 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"runner\" (UniqueName: \"kubernetes.io/empty-dir/5f3f5df9-b819-464d-b7ca-2dea8ceab64d-runner\") pod \"service-telemetry-operator-56dc879589-jp9r2\" (UID: \"5f3f5df9-b819-464d-b7ca-2dea8ceab64d\") " pod="service-telemetry/service-telemetry-operator-56dc879589-jp9r2" Oct 08 00:28:16 crc kubenswrapper[4967]: I1008 00:28:16.653572 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"runner\" (UniqueName: \"kubernetes.io/empty-dir/5f3f5df9-b819-464d-b7ca-2dea8ceab64d-runner\") pod \"service-telemetry-operator-56dc879589-jp9r2\" (UID: \"5f3f5df9-b819-464d-b7ca-2dea8ceab64d\") " pod="service-telemetry/service-telemetry-operator-56dc879589-jp9r2" Oct 08 00:28:16 crc kubenswrapper[4967]: I1008 00:28:16.683595 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sk96r\" (UniqueName: \"kubernetes.io/projected/5f3f5df9-b819-464d-b7ca-2dea8ceab64d-kube-api-access-sk96r\") pod \"service-telemetry-operator-56dc879589-jp9r2\" (UID: \"5f3f5df9-b819-464d-b7ca-2dea8ceab64d\") " pod="service-telemetry/service-telemetry-operator-56dc879589-jp9r2" Oct 08 00:28:16 crc kubenswrapper[4967]: I1008 00:28:16.861020 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-56dc879589-jp9r2" Oct 08 00:28:22 crc kubenswrapper[4967]: I1008 00:28:22.158527 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-56dc879589-jp9r2"] Oct 08 00:28:24 crc kubenswrapper[4967]: W1008 00:28:24.479715 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5f3f5df9_b819_464d_b7ca_2dea8ceab64d.slice/crio-b80e4cfeb1ddb0718b7715e6af772e71baac0237182f9597e73c2dfb970a35d6 WatchSource:0}: Error finding container b80e4cfeb1ddb0718b7715e6af772e71baac0237182f9597e73c2dfb970a35d6: Status 404 returned error can't find the container with id b80e4cfeb1ddb0718b7715e6af772e71baac0237182f9597e73c2dfb970a35d6 Oct 08 00:28:24 crc kubenswrapper[4967]: E1008 00:28:24.818437 4967 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/infrawatch/smart-gateway-operator:stable-1.5" Oct 08 00:28:24 crc kubenswrapper[4967]: E1008 00:28:24.818816 4967 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/infrawatch/smart-gateway-operator:stable-1.5,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:WATCH_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.annotations['olm.targetNamespaces'],},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:OPERATOR_NAME,Value:smart-gateway-operator,ValueFrom:nil,},EnvVar{Name:ANSIBLE_GATHERING,Value:explicit,ValueFrom:nil,},EnvVar{Name:ANSIBLE_VERBOSITY_SMARTGATEWAY_SMARTGATEWAY_INFRA_WATCH,Value:4,ValueFrom:nil,},EnvVar{Name:ANSIBLE_DEBUG_LOGS,Value:true,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CORE_SMARTGATEWAY_IMAGE,Value:image-registry.openshift-image-registry.svc:5000/service-telemetry/sg-core:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BRIDGE_SMARTGATEWAY_IMAGE,Value:image-registry.openshift-image-registry.svc:5000/service-telemetry/sg-bridge:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OAUTH_PROXY_IMAGE,Value:quay.io/openshift/origin-oauth-proxy:latest,ValueFrom:nil,},EnvVar{Name:OPERATOR_CONDITION_NAME,Value:smart-gateway-operator.v5.0.1759883286,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:runner,ReadOnly:false,MountPath:/tmp/ansible-operator/runner,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-c4n4r,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000670000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod smart-gateway-operator-7d98fcf998-q8f6v_service-telemetry(5bc8d2b2-4afa-44de-a704-abacdd934b28): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 08 00:28:24 crc kubenswrapper[4967]: E1008 00:28:24.820137 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="service-telemetry/smart-gateway-operator-7d98fcf998-q8f6v" podUID="5bc8d2b2-4afa-44de-a704-abacdd934b28" Oct 08 00:28:25 crc kubenswrapper[4967]: I1008 00:28:25.400279 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-56dc879589-jp9r2" event={"ID":"5f3f5df9-b819-464d-b7ca-2dea8ceab64d","Type":"ContainerStarted","Data":"b80e4cfeb1ddb0718b7715e6af772e71baac0237182f9597e73c2dfb970a35d6"} Oct 08 00:28:25 crc kubenswrapper[4967]: E1008 00:28:25.402843 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/infrawatch/smart-gateway-operator:stable-1.5\\\"\"" pod="service-telemetry/smart-gateway-operator-7d98fcf998-q8f6v" podUID="5bc8d2b2-4afa-44de-a704-abacdd934b28" Oct 08 00:28:30 crc kubenswrapper[4967]: I1008 00:28:30.435994 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-56dc879589-jp9r2" event={"ID":"5f3f5df9-b819-464d-b7ca-2dea8ceab64d","Type":"ContainerStarted","Data":"4e9b6f065295edc6026a17bf3720349130df231cdfc77a5b63377f822a3ba372"} Oct 08 00:28:30 crc kubenswrapper[4967]: I1008 00:28:30.463226 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/service-telemetry-operator-56dc879589-jp9r2" podStartSLOduration=8.667692573 podStartE2EDuration="14.463209609s" podCreationTimestamp="2025-10-08 00:28:16 +0000 UTC" firstStartedPulling="2025-10-08 00:28:24.483757218 +0000 UTC m=+1303.134291867" lastFinishedPulling="2025-10-08 00:28:30.279274254 +0000 UTC m=+1308.929808903" observedRunningTime="2025-10-08 00:28:30.46127654 +0000 UTC m=+1309.111811269" watchObservedRunningTime="2025-10-08 00:28:30.463209609 +0000 UTC m=+1309.113744258" Oct 08 00:28:39 crc kubenswrapper[4967]: I1008 00:28:39.505998 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-7d98fcf998-q8f6v" event={"ID":"5bc8d2b2-4afa-44de-a704-abacdd934b28","Type":"ContainerStarted","Data":"81ac17a1d6cc796b473de450a5dc79f43e493d0abeac4100d39a800d467d53ac"} Oct 08 00:28:39 crc kubenswrapper[4967]: I1008 00:28:39.535519 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/smart-gateway-operator-7d98fcf998-q8f6v" podStartSLOduration=1.896147826 podStartE2EDuration="29.535490915s" podCreationTimestamp="2025-10-08 00:28:10 +0000 UTC" firstStartedPulling="2025-10-08 00:28:10.931063657 +0000 UTC m=+1289.581598326" lastFinishedPulling="2025-10-08 00:28:38.570406746 +0000 UTC m=+1317.220941415" observedRunningTime="2025-10-08 00:28:39.531393692 +0000 UTC m=+1318.181928351" watchObservedRunningTime="2025-10-08 00:28:39.535490915 +0000 UTC m=+1318.186025574" Oct 08 00:28:54 crc kubenswrapper[4967]: I1008 00:28:54.289801 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/default-interconnect-68864d46cb-xxx8l"] Oct 08 00:28:54 crc kubenswrapper[4967]: I1008 00:28:54.291201 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-interconnect-68864d46cb-xxx8l" Oct 08 00:28:54 crc kubenswrapper[4967]: I1008 00:28:54.297522 4967 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-dockercfg-q28df" Oct 08 00:28:54 crc kubenswrapper[4967]: I1008 00:28:54.298008 4967 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-users" Oct 08 00:28:54 crc kubenswrapper[4967]: I1008 00:28:54.298615 4967 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-openstack-credentials" Oct 08 00:28:54 crc kubenswrapper[4967]: I1008 00:28:54.298812 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"default-interconnect-sasl-config" Oct 08 00:28:54 crc kubenswrapper[4967]: I1008 00:28:54.298957 4967 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-openstack-ca" Oct 08 00:28:54 crc kubenswrapper[4967]: I1008 00:28:54.299061 4967 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-inter-router-ca" Oct 08 00:28:54 crc kubenswrapper[4967]: I1008 00:28:54.299278 4967 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-inter-router-credentials" Oct 08 00:28:54 crc kubenswrapper[4967]: I1008 00:28:54.319554 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-interconnect-68864d46cb-xxx8l"] Oct 08 00:28:54 crc kubenswrapper[4967]: I1008 00:28:54.407932 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-llgwb\" (UniqueName: \"kubernetes.io/projected/e65fe86f-bcf1-41ac-b403-cc48da12ddb8-kube-api-access-llgwb\") pod \"default-interconnect-68864d46cb-xxx8l\" (UID: \"e65fe86f-bcf1-41ac-b403-cc48da12ddb8\") " pod="service-telemetry/default-interconnect-68864d46cb-xxx8l" Oct 08 00:28:54 crc kubenswrapper[4967]: I1008 00:28:54.407998 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/e65fe86f-bcf1-41ac-b403-cc48da12ddb8-default-interconnect-inter-router-credentials\") pod \"default-interconnect-68864d46cb-xxx8l\" (UID: \"e65fe86f-bcf1-41ac-b403-cc48da12ddb8\") " pod="service-telemetry/default-interconnect-68864d46cb-xxx8l" Oct 08 00:28:54 crc kubenswrapper[4967]: I1008 00:28:54.408020 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/e65fe86f-bcf1-41ac-b403-cc48da12ddb8-sasl-config\") pod \"default-interconnect-68864d46cb-xxx8l\" (UID: \"e65fe86f-bcf1-41ac-b403-cc48da12ddb8\") " pod="service-telemetry/default-interconnect-68864d46cb-xxx8l" Oct 08 00:28:54 crc kubenswrapper[4967]: I1008 00:28:54.408089 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/e65fe86f-bcf1-41ac-b403-cc48da12ddb8-default-interconnect-openstack-ca\") pod \"default-interconnect-68864d46cb-xxx8l\" (UID: \"e65fe86f-bcf1-41ac-b403-cc48da12ddb8\") " pod="service-telemetry/default-interconnect-68864d46cb-xxx8l" Oct 08 00:28:54 crc kubenswrapper[4967]: I1008 00:28:54.408111 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/e65fe86f-bcf1-41ac-b403-cc48da12ddb8-sasl-users\") pod \"default-interconnect-68864d46cb-xxx8l\" (UID: \"e65fe86f-bcf1-41ac-b403-cc48da12ddb8\") " pod="service-telemetry/default-interconnect-68864d46cb-xxx8l" Oct 08 00:28:54 crc kubenswrapper[4967]: I1008 00:28:54.408140 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/e65fe86f-bcf1-41ac-b403-cc48da12ddb8-default-interconnect-openstack-credentials\") pod \"default-interconnect-68864d46cb-xxx8l\" (UID: \"e65fe86f-bcf1-41ac-b403-cc48da12ddb8\") " pod="service-telemetry/default-interconnect-68864d46cb-xxx8l" Oct 08 00:28:54 crc kubenswrapper[4967]: I1008 00:28:54.408164 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/e65fe86f-bcf1-41ac-b403-cc48da12ddb8-default-interconnect-inter-router-ca\") pod \"default-interconnect-68864d46cb-xxx8l\" (UID: \"e65fe86f-bcf1-41ac-b403-cc48da12ddb8\") " pod="service-telemetry/default-interconnect-68864d46cb-xxx8l" Oct 08 00:28:54 crc kubenswrapper[4967]: I1008 00:28:54.509366 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/e65fe86f-bcf1-41ac-b403-cc48da12ddb8-default-interconnect-openstack-ca\") pod \"default-interconnect-68864d46cb-xxx8l\" (UID: \"e65fe86f-bcf1-41ac-b403-cc48da12ddb8\") " pod="service-telemetry/default-interconnect-68864d46cb-xxx8l" Oct 08 00:28:54 crc kubenswrapper[4967]: I1008 00:28:54.509407 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/e65fe86f-bcf1-41ac-b403-cc48da12ddb8-sasl-users\") pod \"default-interconnect-68864d46cb-xxx8l\" (UID: \"e65fe86f-bcf1-41ac-b403-cc48da12ddb8\") " pod="service-telemetry/default-interconnect-68864d46cb-xxx8l" Oct 08 00:28:54 crc kubenswrapper[4967]: I1008 00:28:54.509432 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/e65fe86f-bcf1-41ac-b403-cc48da12ddb8-default-interconnect-openstack-credentials\") pod \"default-interconnect-68864d46cb-xxx8l\" (UID: \"e65fe86f-bcf1-41ac-b403-cc48da12ddb8\") " pod="service-telemetry/default-interconnect-68864d46cb-xxx8l" Oct 08 00:28:54 crc kubenswrapper[4967]: I1008 00:28:54.509454 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/e65fe86f-bcf1-41ac-b403-cc48da12ddb8-default-interconnect-inter-router-ca\") pod \"default-interconnect-68864d46cb-xxx8l\" (UID: \"e65fe86f-bcf1-41ac-b403-cc48da12ddb8\") " pod="service-telemetry/default-interconnect-68864d46cb-xxx8l" Oct 08 00:28:54 crc kubenswrapper[4967]: I1008 00:28:54.509490 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-llgwb\" (UniqueName: \"kubernetes.io/projected/e65fe86f-bcf1-41ac-b403-cc48da12ddb8-kube-api-access-llgwb\") pod \"default-interconnect-68864d46cb-xxx8l\" (UID: \"e65fe86f-bcf1-41ac-b403-cc48da12ddb8\") " pod="service-telemetry/default-interconnect-68864d46cb-xxx8l" Oct 08 00:28:54 crc kubenswrapper[4967]: I1008 00:28:54.509524 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/e65fe86f-bcf1-41ac-b403-cc48da12ddb8-default-interconnect-inter-router-credentials\") pod \"default-interconnect-68864d46cb-xxx8l\" (UID: \"e65fe86f-bcf1-41ac-b403-cc48da12ddb8\") " pod="service-telemetry/default-interconnect-68864d46cb-xxx8l" Oct 08 00:28:54 crc kubenswrapper[4967]: I1008 00:28:54.509544 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/e65fe86f-bcf1-41ac-b403-cc48da12ddb8-sasl-config\") pod \"default-interconnect-68864d46cb-xxx8l\" (UID: \"e65fe86f-bcf1-41ac-b403-cc48da12ddb8\") " pod="service-telemetry/default-interconnect-68864d46cb-xxx8l" Oct 08 00:28:54 crc kubenswrapper[4967]: I1008 00:28:54.510527 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/e65fe86f-bcf1-41ac-b403-cc48da12ddb8-sasl-config\") pod \"default-interconnect-68864d46cb-xxx8l\" (UID: \"e65fe86f-bcf1-41ac-b403-cc48da12ddb8\") " pod="service-telemetry/default-interconnect-68864d46cb-xxx8l" Oct 08 00:28:54 crc kubenswrapper[4967]: I1008 00:28:54.514981 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/e65fe86f-bcf1-41ac-b403-cc48da12ddb8-default-interconnect-openstack-credentials\") pod \"default-interconnect-68864d46cb-xxx8l\" (UID: \"e65fe86f-bcf1-41ac-b403-cc48da12ddb8\") " pod="service-telemetry/default-interconnect-68864d46cb-xxx8l" Oct 08 00:28:54 crc kubenswrapper[4967]: I1008 00:28:54.516241 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/e65fe86f-bcf1-41ac-b403-cc48da12ddb8-default-interconnect-inter-router-credentials\") pod \"default-interconnect-68864d46cb-xxx8l\" (UID: \"e65fe86f-bcf1-41ac-b403-cc48da12ddb8\") " pod="service-telemetry/default-interconnect-68864d46cb-xxx8l" Oct 08 00:28:54 crc kubenswrapper[4967]: I1008 00:28:54.517244 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/e65fe86f-bcf1-41ac-b403-cc48da12ddb8-default-interconnect-openstack-ca\") pod \"default-interconnect-68864d46cb-xxx8l\" (UID: \"e65fe86f-bcf1-41ac-b403-cc48da12ddb8\") " pod="service-telemetry/default-interconnect-68864d46cb-xxx8l" Oct 08 00:28:54 crc kubenswrapper[4967]: I1008 00:28:54.523299 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/e65fe86f-bcf1-41ac-b403-cc48da12ddb8-default-interconnect-inter-router-ca\") pod \"default-interconnect-68864d46cb-xxx8l\" (UID: \"e65fe86f-bcf1-41ac-b403-cc48da12ddb8\") " pod="service-telemetry/default-interconnect-68864d46cb-xxx8l" Oct 08 00:28:54 crc kubenswrapper[4967]: I1008 00:28:54.524070 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/e65fe86f-bcf1-41ac-b403-cc48da12ddb8-sasl-users\") pod \"default-interconnect-68864d46cb-xxx8l\" (UID: \"e65fe86f-bcf1-41ac-b403-cc48da12ddb8\") " pod="service-telemetry/default-interconnect-68864d46cb-xxx8l" Oct 08 00:28:54 crc kubenswrapper[4967]: I1008 00:28:54.526109 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-llgwb\" (UniqueName: \"kubernetes.io/projected/e65fe86f-bcf1-41ac-b403-cc48da12ddb8-kube-api-access-llgwb\") pod \"default-interconnect-68864d46cb-xxx8l\" (UID: \"e65fe86f-bcf1-41ac-b403-cc48da12ddb8\") " pod="service-telemetry/default-interconnect-68864d46cb-xxx8l" Oct 08 00:28:54 crc kubenswrapper[4967]: I1008 00:28:54.617115 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-interconnect-68864d46cb-xxx8l" Oct 08 00:28:55 crc kubenswrapper[4967]: I1008 00:28:55.114005 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-interconnect-68864d46cb-xxx8l"] Oct 08 00:28:55 crc kubenswrapper[4967]: I1008 00:28:55.645449 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-interconnect-68864d46cb-xxx8l" event={"ID":"e65fe86f-bcf1-41ac-b403-cc48da12ddb8","Type":"ContainerStarted","Data":"061646feae812d3f37b1837e7d72df04d4d30adffc838c77a08d0277a1c8f863"} Oct 08 00:29:02 crc kubenswrapper[4967]: I1008 00:29:02.697114 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-interconnect-68864d46cb-xxx8l" event={"ID":"e65fe86f-bcf1-41ac-b403-cc48da12ddb8","Type":"ContainerStarted","Data":"34e4a19674abb9a42557c3c696e1ec1784a807abe6c44aa3f5afdced1032a0c6"} Oct 08 00:29:02 crc kubenswrapper[4967]: I1008 00:29:02.726000 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-interconnect-68864d46cb-xxx8l" podStartSLOduration=1.367136625 podStartE2EDuration="8.725927313s" podCreationTimestamp="2025-10-08 00:28:54 +0000 UTC" firstStartedPulling="2025-10-08 00:28:55.121865987 +0000 UTC m=+1333.772400646" lastFinishedPulling="2025-10-08 00:29:02.480656685 +0000 UTC m=+1341.131191334" observedRunningTime="2025-10-08 00:29:02.717728637 +0000 UTC m=+1341.368263316" watchObservedRunningTime="2025-10-08 00:29:02.725927313 +0000 UTC m=+1341.376462002" Oct 08 00:29:04 crc kubenswrapper[4967]: I1008 00:29:04.480765 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/prometheus-default-0"] Oct 08 00:29:04 crc kubenswrapper[4967]: I1008 00:29:04.483599 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-default-0" Oct 08 00:29:04 crc kubenswrapper[4967]: I1008 00:29:04.486974 4967 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-session-secret" Oct 08 00:29:04 crc kubenswrapper[4967]: I1008 00:29:04.487551 4967 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"prometheus-default-tls-assets-0" Oct 08 00:29:04 crc kubenswrapper[4967]: I1008 00:29:04.489334 4967 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"prometheus-stf-dockercfg-l5zsr" Oct 08 00:29:04 crc kubenswrapper[4967]: I1008 00:29:04.490142 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"prometheus-default-rulefiles-0" Oct 08 00:29:04 crc kubenswrapper[4967]: I1008 00:29:04.490805 4967 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-prometheus-proxy-tls" Oct 08 00:29:04 crc kubenswrapper[4967]: I1008 00:29:04.491165 4967 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"prometheus-default" Oct 08 00:29:04 crc kubenswrapper[4967]: I1008 00:29:04.491190 4967 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"prometheus-default-web-config" Oct 08 00:29:04 crc kubenswrapper[4967]: I1008 00:29:04.492158 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"serving-certs-ca-bundle" Oct 08 00:29:04 crc kubenswrapper[4967]: I1008 00:29:04.500856 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/prometheus-default-0"] Oct 08 00:29:04 crc kubenswrapper[4967]: I1008 00:29:04.665656 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-default-session-secret\" (UniqueName: \"kubernetes.io/secret/82ce08ad-02fd-4ae6-b3c5-c2c01ac6fdec-secret-default-session-secret\") pod \"prometheus-default-0\" (UID: \"82ce08ad-02fd-4ae6-b3c5-c2c01ac6fdec\") " pod="service-telemetry/prometheus-default-0" Oct 08 00:29:04 crc kubenswrapper[4967]: I1008 00:29:04.666040 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/82ce08ad-02fd-4ae6-b3c5-c2c01ac6fdec-config\") pod \"prometheus-default-0\" (UID: \"82ce08ad-02fd-4ae6-b3c5-c2c01ac6fdec\") " pod="service-telemetry/prometheus-default-0" Oct 08 00:29:04 crc kubenswrapper[4967]: I1008 00:29:04.666196 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/82ce08ad-02fd-4ae6-b3c5-c2c01ac6fdec-configmap-serving-certs-ca-bundle\") pod \"prometheus-default-0\" (UID: \"82ce08ad-02fd-4ae6-b3c5-c2c01ac6fdec\") " pod="service-telemetry/prometheus-default-0" Oct 08 00:29:04 crc kubenswrapper[4967]: I1008 00:29:04.666344 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-default-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/82ce08ad-02fd-4ae6-b3c5-c2c01ac6fdec-prometheus-default-rulefiles-0\") pod \"prometheus-default-0\" (UID: \"82ce08ad-02fd-4ae6-b3c5-c2c01ac6fdec\") " pod="service-telemetry/prometheus-default-0" Oct 08 00:29:04 crc kubenswrapper[4967]: I1008 00:29:04.666398 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/82ce08ad-02fd-4ae6-b3c5-c2c01ac6fdec-tls-assets\") pod \"prometheus-default-0\" (UID: \"82ce08ad-02fd-4ae6-b3c5-c2c01ac6fdec\") " pod="service-telemetry/prometheus-default-0" Oct 08 00:29:04 crc kubenswrapper[4967]: I1008 00:29:04.666489 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-62eee05b-8537-439f-a43b-02a0360ef62e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-62eee05b-8537-439f-a43b-02a0360ef62e\") pod \"prometheus-default-0\" (UID: \"82ce08ad-02fd-4ae6-b3c5-c2c01ac6fdec\") " pod="service-telemetry/prometheus-default-0" Oct 08 00:29:04 crc kubenswrapper[4967]: I1008 00:29:04.666535 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/82ce08ad-02fd-4ae6-b3c5-c2c01ac6fdec-web-config\") pod \"prometheus-default-0\" (UID: \"82ce08ad-02fd-4ae6-b3c5-c2c01ac6fdec\") " pod="service-telemetry/prometheus-default-0" Oct 08 00:29:04 crc kubenswrapper[4967]: I1008 00:29:04.666569 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-default-prometheus-proxy-tls\" (UniqueName: \"kubernetes.io/secret/82ce08ad-02fd-4ae6-b3c5-c2c01ac6fdec-secret-default-prometheus-proxy-tls\") pod \"prometheus-default-0\" (UID: \"82ce08ad-02fd-4ae6-b3c5-c2c01ac6fdec\") " pod="service-telemetry/prometheus-default-0" Oct 08 00:29:04 crc kubenswrapper[4967]: I1008 00:29:04.666596 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zq5dx\" (UniqueName: \"kubernetes.io/projected/82ce08ad-02fd-4ae6-b3c5-c2c01ac6fdec-kube-api-access-zq5dx\") pod \"prometheus-default-0\" (UID: \"82ce08ad-02fd-4ae6-b3c5-c2c01ac6fdec\") " pod="service-telemetry/prometheus-default-0" Oct 08 00:29:04 crc kubenswrapper[4967]: I1008 00:29:04.666626 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/82ce08ad-02fd-4ae6-b3c5-c2c01ac6fdec-config-out\") pod \"prometheus-default-0\" (UID: \"82ce08ad-02fd-4ae6-b3c5-c2c01ac6fdec\") " pod="service-telemetry/prometheus-default-0" Oct 08 00:29:04 crc kubenswrapper[4967]: I1008 00:29:04.768080 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/82ce08ad-02fd-4ae6-b3c5-c2c01ac6fdec-configmap-serving-certs-ca-bundle\") pod \"prometheus-default-0\" (UID: \"82ce08ad-02fd-4ae6-b3c5-c2c01ac6fdec\") " pod="service-telemetry/prometheus-default-0" Oct 08 00:29:04 crc kubenswrapper[4967]: I1008 00:29:04.768135 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-default-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/82ce08ad-02fd-4ae6-b3c5-c2c01ac6fdec-prometheus-default-rulefiles-0\") pod \"prometheus-default-0\" (UID: \"82ce08ad-02fd-4ae6-b3c5-c2c01ac6fdec\") " pod="service-telemetry/prometheus-default-0" Oct 08 00:29:04 crc kubenswrapper[4967]: I1008 00:29:04.768157 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/82ce08ad-02fd-4ae6-b3c5-c2c01ac6fdec-tls-assets\") pod \"prometheus-default-0\" (UID: \"82ce08ad-02fd-4ae6-b3c5-c2c01ac6fdec\") " pod="service-telemetry/prometheus-default-0" Oct 08 00:29:04 crc kubenswrapper[4967]: I1008 00:29:04.768244 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-62eee05b-8537-439f-a43b-02a0360ef62e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-62eee05b-8537-439f-a43b-02a0360ef62e\") pod \"prometheus-default-0\" (UID: \"82ce08ad-02fd-4ae6-b3c5-c2c01ac6fdec\") " pod="service-telemetry/prometheus-default-0" Oct 08 00:29:04 crc kubenswrapper[4967]: I1008 00:29:04.768280 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/82ce08ad-02fd-4ae6-b3c5-c2c01ac6fdec-web-config\") pod \"prometheus-default-0\" (UID: \"82ce08ad-02fd-4ae6-b3c5-c2c01ac6fdec\") " pod="service-telemetry/prometheus-default-0" Oct 08 00:29:04 crc kubenswrapper[4967]: I1008 00:29:04.768305 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-default-prometheus-proxy-tls\" (UniqueName: \"kubernetes.io/secret/82ce08ad-02fd-4ae6-b3c5-c2c01ac6fdec-secret-default-prometheus-proxy-tls\") pod \"prometheus-default-0\" (UID: \"82ce08ad-02fd-4ae6-b3c5-c2c01ac6fdec\") " pod="service-telemetry/prometheus-default-0" Oct 08 00:29:04 crc kubenswrapper[4967]: I1008 00:29:04.768328 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zq5dx\" (UniqueName: \"kubernetes.io/projected/82ce08ad-02fd-4ae6-b3c5-c2c01ac6fdec-kube-api-access-zq5dx\") pod \"prometheus-default-0\" (UID: \"82ce08ad-02fd-4ae6-b3c5-c2c01ac6fdec\") " pod="service-telemetry/prometheus-default-0" Oct 08 00:29:04 crc kubenswrapper[4967]: I1008 00:29:04.768348 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/82ce08ad-02fd-4ae6-b3c5-c2c01ac6fdec-config-out\") pod \"prometheus-default-0\" (UID: \"82ce08ad-02fd-4ae6-b3c5-c2c01ac6fdec\") " pod="service-telemetry/prometheus-default-0" Oct 08 00:29:04 crc kubenswrapper[4967]: I1008 00:29:04.768370 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-default-session-secret\" (UniqueName: \"kubernetes.io/secret/82ce08ad-02fd-4ae6-b3c5-c2c01ac6fdec-secret-default-session-secret\") pod \"prometheus-default-0\" (UID: \"82ce08ad-02fd-4ae6-b3c5-c2c01ac6fdec\") " pod="service-telemetry/prometheus-default-0" Oct 08 00:29:04 crc kubenswrapper[4967]: I1008 00:29:04.768390 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/82ce08ad-02fd-4ae6-b3c5-c2c01ac6fdec-config\") pod \"prometheus-default-0\" (UID: \"82ce08ad-02fd-4ae6-b3c5-c2c01ac6fdec\") " pod="service-telemetry/prometheus-default-0" Oct 08 00:29:04 crc kubenswrapper[4967]: I1008 00:29:04.771172 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/82ce08ad-02fd-4ae6-b3c5-c2c01ac6fdec-configmap-serving-certs-ca-bundle\") pod \"prometheus-default-0\" (UID: \"82ce08ad-02fd-4ae6-b3c5-c2c01ac6fdec\") " pod="service-telemetry/prometheus-default-0" Oct 08 00:29:04 crc kubenswrapper[4967]: I1008 00:29:04.771324 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-default-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/82ce08ad-02fd-4ae6-b3c5-c2c01ac6fdec-prometheus-default-rulefiles-0\") pod \"prometheus-default-0\" (UID: \"82ce08ad-02fd-4ae6-b3c5-c2c01ac6fdec\") " pod="service-telemetry/prometheus-default-0" Oct 08 00:29:04 crc kubenswrapper[4967]: E1008 00:29:04.771498 4967 secret.go:188] Couldn't get secret service-telemetry/default-prometheus-proxy-tls: secret "default-prometheus-proxy-tls" not found Oct 08 00:29:04 crc kubenswrapper[4967]: E1008 00:29:04.771556 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/82ce08ad-02fd-4ae6-b3c5-c2c01ac6fdec-secret-default-prometheus-proxy-tls podName:82ce08ad-02fd-4ae6-b3c5-c2c01ac6fdec nodeName:}" failed. No retries permitted until 2025-10-08 00:29:05.271539753 +0000 UTC m=+1343.922074402 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "secret-default-prometheus-proxy-tls" (UniqueName: "kubernetes.io/secret/82ce08ad-02fd-4ae6-b3c5-c2c01ac6fdec-secret-default-prometheus-proxy-tls") pod "prometheus-default-0" (UID: "82ce08ad-02fd-4ae6-b3c5-c2c01ac6fdec") : secret "default-prometheus-proxy-tls" not found Oct 08 00:29:04 crc kubenswrapper[4967]: I1008 00:29:04.776208 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/82ce08ad-02fd-4ae6-b3c5-c2c01ac6fdec-tls-assets\") pod \"prometheus-default-0\" (UID: \"82ce08ad-02fd-4ae6-b3c5-c2c01ac6fdec\") " pod="service-telemetry/prometheus-default-0" Oct 08 00:29:04 crc kubenswrapper[4967]: I1008 00:29:04.777284 4967 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 08 00:29:04 crc kubenswrapper[4967]: I1008 00:29:04.777984 4967 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-62eee05b-8537-439f-a43b-02a0360ef62e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-62eee05b-8537-439f-a43b-02a0360ef62e\") pod \"prometheus-default-0\" (UID: \"82ce08ad-02fd-4ae6-b3c5-c2c01ac6fdec\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/5708fdf23681ae574f1c1ba202019c2f4ec3a764bd9dc1a8c4716263fa5acff4/globalmount\"" pod="service-telemetry/prometheus-default-0" Oct 08 00:29:04 crc kubenswrapper[4967]: I1008 00:29:04.781997 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/82ce08ad-02fd-4ae6-b3c5-c2c01ac6fdec-config-out\") pod \"prometheus-default-0\" (UID: \"82ce08ad-02fd-4ae6-b3c5-c2c01ac6fdec\") " pod="service-telemetry/prometheus-default-0" Oct 08 00:29:04 crc kubenswrapper[4967]: I1008 00:29:04.782538 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-default-session-secret\" (UniqueName: \"kubernetes.io/secret/82ce08ad-02fd-4ae6-b3c5-c2c01ac6fdec-secret-default-session-secret\") pod \"prometheus-default-0\" (UID: \"82ce08ad-02fd-4ae6-b3c5-c2c01ac6fdec\") " pod="service-telemetry/prometheus-default-0" Oct 08 00:29:04 crc kubenswrapper[4967]: I1008 00:29:04.785294 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/82ce08ad-02fd-4ae6-b3c5-c2c01ac6fdec-web-config\") pod \"prometheus-default-0\" (UID: \"82ce08ad-02fd-4ae6-b3c5-c2c01ac6fdec\") " pod="service-telemetry/prometheus-default-0" Oct 08 00:29:04 crc kubenswrapper[4967]: I1008 00:29:04.787131 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/82ce08ad-02fd-4ae6-b3c5-c2c01ac6fdec-config\") pod \"prometheus-default-0\" (UID: \"82ce08ad-02fd-4ae6-b3c5-c2c01ac6fdec\") " pod="service-telemetry/prometheus-default-0" Oct 08 00:29:04 crc kubenswrapper[4967]: I1008 00:29:04.808393 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-62eee05b-8537-439f-a43b-02a0360ef62e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-62eee05b-8537-439f-a43b-02a0360ef62e\") pod \"prometheus-default-0\" (UID: \"82ce08ad-02fd-4ae6-b3c5-c2c01ac6fdec\") " pod="service-telemetry/prometheus-default-0" Oct 08 00:29:04 crc kubenswrapper[4967]: I1008 00:29:04.810050 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zq5dx\" (UniqueName: \"kubernetes.io/projected/82ce08ad-02fd-4ae6-b3c5-c2c01ac6fdec-kube-api-access-zq5dx\") pod \"prometheus-default-0\" (UID: \"82ce08ad-02fd-4ae6-b3c5-c2c01ac6fdec\") " pod="service-telemetry/prometheus-default-0" Oct 08 00:29:05 crc kubenswrapper[4967]: I1008 00:29:05.274568 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-default-prometheus-proxy-tls\" (UniqueName: \"kubernetes.io/secret/82ce08ad-02fd-4ae6-b3c5-c2c01ac6fdec-secret-default-prometheus-proxy-tls\") pod \"prometheus-default-0\" (UID: \"82ce08ad-02fd-4ae6-b3c5-c2c01ac6fdec\") " pod="service-telemetry/prometheus-default-0" Oct 08 00:29:05 crc kubenswrapper[4967]: E1008 00:29:05.275696 4967 secret.go:188] Couldn't get secret service-telemetry/default-prometheus-proxy-tls: secret "default-prometheus-proxy-tls" not found Oct 08 00:29:05 crc kubenswrapper[4967]: E1008 00:29:05.275784 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/82ce08ad-02fd-4ae6-b3c5-c2c01ac6fdec-secret-default-prometheus-proxy-tls podName:82ce08ad-02fd-4ae6-b3c5-c2c01ac6fdec nodeName:}" failed. No retries permitted until 2025-10-08 00:29:06.275764432 +0000 UTC m=+1344.926299091 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "secret-default-prometheus-proxy-tls" (UniqueName: "kubernetes.io/secret/82ce08ad-02fd-4ae6-b3c5-c2c01ac6fdec-secret-default-prometheus-proxy-tls") pod "prometheus-default-0" (UID: "82ce08ad-02fd-4ae6-b3c5-c2c01ac6fdec") : secret "default-prometheus-proxy-tls" not found Oct 08 00:29:06 crc kubenswrapper[4967]: I1008 00:29:06.296680 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-default-prometheus-proxy-tls\" (UniqueName: \"kubernetes.io/secret/82ce08ad-02fd-4ae6-b3c5-c2c01ac6fdec-secret-default-prometheus-proxy-tls\") pod \"prometheus-default-0\" (UID: \"82ce08ad-02fd-4ae6-b3c5-c2c01ac6fdec\") " pod="service-telemetry/prometheus-default-0" Oct 08 00:29:06 crc kubenswrapper[4967]: I1008 00:29:06.313372 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-default-prometheus-proxy-tls\" (UniqueName: \"kubernetes.io/secret/82ce08ad-02fd-4ae6-b3c5-c2c01ac6fdec-secret-default-prometheus-proxy-tls\") pod \"prometheus-default-0\" (UID: \"82ce08ad-02fd-4ae6-b3c5-c2c01ac6fdec\") " pod="service-telemetry/prometheus-default-0" Oct 08 00:29:06 crc kubenswrapper[4967]: I1008 00:29:06.605645 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-default-0" Oct 08 00:29:07 crc kubenswrapper[4967]: I1008 00:29:07.050884 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/prometheus-default-0"] Oct 08 00:29:07 crc kubenswrapper[4967]: W1008 00:29:07.053280 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod82ce08ad_02fd_4ae6_b3c5_c2c01ac6fdec.slice/crio-deb8b273cd305fe76a7820c7fb240c17f0c698dec1cbd8be3b994f9932ca4c1d WatchSource:0}: Error finding container deb8b273cd305fe76a7820c7fb240c17f0c698dec1cbd8be3b994f9932ca4c1d: Status 404 returned error can't find the container with id deb8b273cd305fe76a7820c7fb240c17f0c698dec1cbd8be3b994f9932ca4c1d Oct 08 00:29:07 crc kubenswrapper[4967]: I1008 00:29:07.736890 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-default-0" event={"ID":"82ce08ad-02fd-4ae6-b3c5-c2c01ac6fdec","Type":"ContainerStarted","Data":"deb8b273cd305fe76a7820c7fb240c17f0c698dec1cbd8be3b994f9932ca4c1d"} Oct 08 00:29:11 crc kubenswrapper[4967]: I1008 00:29:11.767152 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-default-0" event={"ID":"82ce08ad-02fd-4ae6-b3c5-c2c01ac6fdec","Type":"ContainerStarted","Data":"b67d9ba8ddaa2318500760e6ca0d7ad92cb9f715da966aa161799d7154559ca2"} Oct 08 00:29:14 crc kubenswrapper[4967]: I1008 00:29:14.592248 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/default-snmp-webhook-6856cfb745-6jgkj"] Oct 08 00:29:14 crc kubenswrapper[4967]: I1008 00:29:14.594240 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-snmp-webhook-6856cfb745-6jgkj" Oct 08 00:29:14 crc kubenswrapper[4967]: I1008 00:29:14.605406 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-snmp-webhook-6856cfb745-6jgkj"] Oct 08 00:29:14 crc kubenswrapper[4967]: I1008 00:29:14.640183 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sswqs\" (UniqueName: \"kubernetes.io/projected/f3eb705e-460b-45f4-be46-8d811bda63c3-kube-api-access-sswqs\") pod \"default-snmp-webhook-6856cfb745-6jgkj\" (UID: \"f3eb705e-460b-45f4-be46-8d811bda63c3\") " pod="service-telemetry/default-snmp-webhook-6856cfb745-6jgkj" Oct 08 00:29:14 crc kubenswrapper[4967]: I1008 00:29:14.741300 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sswqs\" (UniqueName: \"kubernetes.io/projected/f3eb705e-460b-45f4-be46-8d811bda63c3-kube-api-access-sswqs\") pod \"default-snmp-webhook-6856cfb745-6jgkj\" (UID: \"f3eb705e-460b-45f4-be46-8d811bda63c3\") " pod="service-telemetry/default-snmp-webhook-6856cfb745-6jgkj" Oct 08 00:29:14 crc kubenswrapper[4967]: I1008 00:29:14.761060 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sswqs\" (UniqueName: \"kubernetes.io/projected/f3eb705e-460b-45f4-be46-8d811bda63c3-kube-api-access-sswqs\") pod \"default-snmp-webhook-6856cfb745-6jgkj\" (UID: \"f3eb705e-460b-45f4-be46-8d811bda63c3\") " pod="service-telemetry/default-snmp-webhook-6856cfb745-6jgkj" Oct 08 00:29:14 crc kubenswrapper[4967]: I1008 00:29:14.912633 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-snmp-webhook-6856cfb745-6jgkj" Oct 08 00:29:15 crc kubenswrapper[4967]: I1008 00:29:15.154261 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-snmp-webhook-6856cfb745-6jgkj"] Oct 08 00:29:15 crc kubenswrapper[4967]: I1008 00:29:15.798787 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-snmp-webhook-6856cfb745-6jgkj" event={"ID":"f3eb705e-460b-45f4-be46-8d811bda63c3","Type":"ContainerStarted","Data":"24949a808092f6bfb95570130f8a94ef82ee068d192523edebf40280df11e244"} Oct 08 00:29:18 crc kubenswrapper[4967]: I1008 00:29:18.446939 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/alertmanager-default-0"] Oct 08 00:29:18 crc kubenswrapper[4967]: I1008 00:29:18.460437 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/alertmanager-default-0" Oct 08 00:29:18 crc kubenswrapper[4967]: I1008 00:29:18.463629 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/alertmanager-default-0"] Oct 08 00:29:18 crc kubenswrapper[4967]: I1008 00:29:18.491819 4967 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"alertmanager-default-tls-assets-0" Oct 08 00:29:18 crc kubenswrapper[4967]: I1008 00:29:18.492710 4967 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-alertmanager-proxy-tls" Oct 08 00:29:18 crc kubenswrapper[4967]: I1008 00:29:18.492828 4967 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"alertmanager-stf-dockercfg-5zkvw" Oct 08 00:29:18 crc kubenswrapper[4967]: I1008 00:29:18.492987 4967 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"alertmanager-default-web-config" Oct 08 00:29:18 crc kubenswrapper[4967]: I1008 00:29:18.493248 4967 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"alertmanager-default-generated" Oct 08 00:29:18 crc kubenswrapper[4967]: I1008 00:29:18.600550 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/9ce411af-8da3-4002-a75b-bc8ba4a64db8-tls-assets\") pod \"alertmanager-default-0\" (UID: \"9ce411af-8da3-4002-a75b-bc8ba4a64db8\") " pod="service-telemetry/alertmanager-default-0" Oct 08 00:29:18 crc kubenswrapper[4967]: I1008 00:29:18.600609 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-default-session-secret\" (UniqueName: \"kubernetes.io/secret/9ce411af-8da3-4002-a75b-bc8ba4a64db8-secret-default-session-secret\") pod \"alertmanager-default-0\" (UID: \"9ce411af-8da3-4002-a75b-bc8ba4a64db8\") " pod="service-telemetry/alertmanager-default-0" Oct 08 00:29:18 crc kubenswrapper[4967]: I1008 00:29:18.600730 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/9ce411af-8da3-4002-a75b-bc8ba4a64db8-web-config\") pod \"alertmanager-default-0\" (UID: \"9ce411af-8da3-4002-a75b-bc8ba4a64db8\") " pod="service-telemetry/alertmanager-default-0" Oct 08 00:29:18 crc kubenswrapper[4967]: I1008 00:29:18.600781 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/9ce411af-8da3-4002-a75b-bc8ba4a64db8-config-volume\") pod \"alertmanager-default-0\" (UID: \"9ce411af-8da3-4002-a75b-bc8ba4a64db8\") " pod="service-telemetry/alertmanager-default-0" Oct 08 00:29:18 crc kubenswrapper[4967]: I1008 00:29:18.600886 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-546c7d66-8203-4b0e-9504-d1fa4a6cd6c4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-546c7d66-8203-4b0e-9504-d1fa4a6cd6c4\") pod \"alertmanager-default-0\" (UID: \"9ce411af-8da3-4002-a75b-bc8ba4a64db8\") " pod="service-telemetry/alertmanager-default-0" Oct 08 00:29:18 crc kubenswrapper[4967]: I1008 00:29:18.600963 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w96mk\" (UniqueName: \"kubernetes.io/projected/9ce411af-8da3-4002-a75b-bc8ba4a64db8-kube-api-access-w96mk\") pod \"alertmanager-default-0\" (UID: \"9ce411af-8da3-4002-a75b-bc8ba4a64db8\") " pod="service-telemetry/alertmanager-default-0" Oct 08 00:29:18 crc kubenswrapper[4967]: I1008 00:29:18.600994 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/9ce411af-8da3-4002-a75b-bc8ba4a64db8-config-out\") pod \"alertmanager-default-0\" (UID: \"9ce411af-8da3-4002-a75b-bc8ba4a64db8\") " pod="service-telemetry/alertmanager-default-0" Oct 08 00:29:18 crc kubenswrapper[4967]: I1008 00:29:18.601122 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-default-alertmanager-proxy-tls\" (UniqueName: \"kubernetes.io/secret/9ce411af-8da3-4002-a75b-bc8ba4a64db8-secret-default-alertmanager-proxy-tls\") pod \"alertmanager-default-0\" (UID: \"9ce411af-8da3-4002-a75b-bc8ba4a64db8\") " pod="service-telemetry/alertmanager-default-0" Oct 08 00:29:18 crc kubenswrapper[4967]: I1008 00:29:18.702017 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w96mk\" (UniqueName: \"kubernetes.io/projected/9ce411af-8da3-4002-a75b-bc8ba4a64db8-kube-api-access-w96mk\") pod \"alertmanager-default-0\" (UID: \"9ce411af-8da3-4002-a75b-bc8ba4a64db8\") " pod="service-telemetry/alertmanager-default-0" Oct 08 00:29:18 crc kubenswrapper[4967]: I1008 00:29:18.702056 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/9ce411af-8da3-4002-a75b-bc8ba4a64db8-config-out\") pod \"alertmanager-default-0\" (UID: \"9ce411af-8da3-4002-a75b-bc8ba4a64db8\") " pod="service-telemetry/alertmanager-default-0" Oct 08 00:29:18 crc kubenswrapper[4967]: I1008 00:29:18.702115 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-default-alertmanager-proxy-tls\" (UniqueName: \"kubernetes.io/secret/9ce411af-8da3-4002-a75b-bc8ba4a64db8-secret-default-alertmanager-proxy-tls\") pod \"alertmanager-default-0\" (UID: \"9ce411af-8da3-4002-a75b-bc8ba4a64db8\") " pod="service-telemetry/alertmanager-default-0" Oct 08 00:29:18 crc kubenswrapper[4967]: I1008 00:29:18.702155 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/9ce411af-8da3-4002-a75b-bc8ba4a64db8-tls-assets\") pod \"alertmanager-default-0\" (UID: \"9ce411af-8da3-4002-a75b-bc8ba4a64db8\") " pod="service-telemetry/alertmanager-default-0" Oct 08 00:29:18 crc kubenswrapper[4967]: I1008 00:29:18.702357 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-default-session-secret\" (UniqueName: \"kubernetes.io/secret/9ce411af-8da3-4002-a75b-bc8ba4a64db8-secret-default-session-secret\") pod \"alertmanager-default-0\" (UID: \"9ce411af-8da3-4002-a75b-bc8ba4a64db8\") " pod="service-telemetry/alertmanager-default-0" Oct 08 00:29:18 crc kubenswrapper[4967]: I1008 00:29:18.702383 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/9ce411af-8da3-4002-a75b-bc8ba4a64db8-web-config\") pod \"alertmanager-default-0\" (UID: \"9ce411af-8da3-4002-a75b-bc8ba4a64db8\") " pod="service-telemetry/alertmanager-default-0" Oct 08 00:29:18 crc kubenswrapper[4967]: I1008 00:29:18.702403 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/9ce411af-8da3-4002-a75b-bc8ba4a64db8-config-volume\") pod \"alertmanager-default-0\" (UID: \"9ce411af-8da3-4002-a75b-bc8ba4a64db8\") " pod="service-telemetry/alertmanager-default-0" Oct 08 00:29:18 crc kubenswrapper[4967]: I1008 00:29:18.702450 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-546c7d66-8203-4b0e-9504-d1fa4a6cd6c4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-546c7d66-8203-4b0e-9504-d1fa4a6cd6c4\") pod \"alertmanager-default-0\" (UID: \"9ce411af-8da3-4002-a75b-bc8ba4a64db8\") " pod="service-telemetry/alertmanager-default-0" Oct 08 00:29:18 crc kubenswrapper[4967]: E1008 00:29:18.702576 4967 secret.go:188] Couldn't get secret service-telemetry/default-alertmanager-proxy-tls: secret "default-alertmanager-proxy-tls" not found Oct 08 00:29:18 crc kubenswrapper[4967]: E1008 00:29:18.702710 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9ce411af-8da3-4002-a75b-bc8ba4a64db8-secret-default-alertmanager-proxy-tls podName:9ce411af-8da3-4002-a75b-bc8ba4a64db8 nodeName:}" failed. No retries permitted until 2025-10-08 00:29:19.202652841 +0000 UTC m=+1357.853187560 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "secret-default-alertmanager-proxy-tls" (UniqueName: "kubernetes.io/secret/9ce411af-8da3-4002-a75b-bc8ba4a64db8-secret-default-alertmanager-proxy-tls") pod "alertmanager-default-0" (UID: "9ce411af-8da3-4002-a75b-bc8ba4a64db8") : secret "default-alertmanager-proxy-tls" not found Oct 08 00:29:18 crc kubenswrapper[4967]: I1008 00:29:18.708572 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/9ce411af-8da3-4002-a75b-bc8ba4a64db8-tls-assets\") pod \"alertmanager-default-0\" (UID: \"9ce411af-8da3-4002-a75b-bc8ba4a64db8\") " pod="service-telemetry/alertmanager-default-0" Oct 08 00:29:18 crc kubenswrapper[4967]: I1008 00:29:18.709053 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/9ce411af-8da3-4002-a75b-bc8ba4a64db8-web-config\") pod \"alertmanager-default-0\" (UID: \"9ce411af-8da3-4002-a75b-bc8ba4a64db8\") " pod="service-telemetry/alertmanager-default-0" Oct 08 00:29:18 crc kubenswrapper[4967]: I1008 00:29:18.709343 4967 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 08 00:29:18 crc kubenswrapper[4967]: I1008 00:29:18.715954 4967 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-546c7d66-8203-4b0e-9504-d1fa4a6cd6c4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-546c7d66-8203-4b0e-9504-d1fa4a6cd6c4\") pod \"alertmanager-default-0\" (UID: \"9ce411af-8da3-4002-a75b-bc8ba4a64db8\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/db121662fe938939961a02c1f8249f48f609b88cc148b42a868ae4068d7b6a7a/globalmount\"" pod="service-telemetry/alertmanager-default-0" Oct 08 00:29:18 crc kubenswrapper[4967]: I1008 00:29:18.709391 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/9ce411af-8da3-4002-a75b-bc8ba4a64db8-config-volume\") pod \"alertmanager-default-0\" (UID: \"9ce411af-8da3-4002-a75b-bc8ba4a64db8\") " pod="service-telemetry/alertmanager-default-0" Oct 08 00:29:18 crc kubenswrapper[4967]: I1008 00:29:18.714278 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-default-session-secret\" (UniqueName: \"kubernetes.io/secret/9ce411af-8da3-4002-a75b-bc8ba4a64db8-secret-default-session-secret\") pod \"alertmanager-default-0\" (UID: \"9ce411af-8da3-4002-a75b-bc8ba4a64db8\") " pod="service-telemetry/alertmanager-default-0" Oct 08 00:29:18 crc kubenswrapper[4967]: I1008 00:29:18.709407 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/9ce411af-8da3-4002-a75b-bc8ba4a64db8-config-out\") pod \"alertmanager-default-0\" (UID: \"9ce411af-8da3-4002-a75b-bc8ba4a64db8\") " pod="service-telemetry/alertmanager-default-0" Oct 08 00:29:18 crc kubenswrapper[4967]: I1008 00:29:18.724085 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w96mk\" (UniqueName: \"kubernetes.io/projected/9ce411af-8da3-4002-a75b-bc8ba4a64db8-kube-api-access-w96mk\") pod \"alertmanager-default-0\" (UID: \"9ce411af-8da3-4002-a75b-bc8ba4a64db8\") " pod="service-telemetry/alertmanager-default-0" Oct 08 00:29:18 crc kubenswrapper[4967]: I1008 00:29:18.740464 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-546c7d66-8203-4b0e-9504-d1fa4a6cd6c4\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-546c7d66-8203-4b0e-9504-d1fa4a6cd6c4\") pod \"alertmanager-default-0\" (UID: \"9ce411af-8da3-4002-a75b-bc8ba4a64db8\") " pod="service-telemetry/alertmanager-default-0" Oct 08 00:29:19 crc kubenswrapper[4967]: I1008 00:29:19.209159 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-default-alertmanager-proxy-tls\" (UniqueName: \"kubernetes.io/secret/9ce411af-8da3-4002-a75b-bc8ba4a64db8-secret-default-alertmanager-proxy-tls\") pod \"alertmanager-default-0\" (UID: \"9ce411af-8da3-4002-a75b-bc8ba4a64db8\") " pod="service-telemetry/alertmanager-default-0" Oct 08 00:29:19 crc kubenswrapper[4967]: E1008 00:29:19.209385 4967 secret.go:188] Couldn't get secret service-telemetry/default-alertmanager-proxy-tls: secret "default-alertmanager-proxy-tls" not found Oct 08 00:29:19 crc kubenswrapper[4967]: E1008 00:29:19.209482 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9ce411af-8da3-4002-a75b-bc8ba4a64db8-secret-default-alertmanager-proxy-tls podName:9ce411af-8da3-4002-a75b-bc8ba4a64db8 nodeName:}" failed. No retries permitted until 2025-10-08 00:29:20.209458886 +0000 UTC m=+1358.859993535 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "secret-default-alertmanager-proxy-tls" (UniqueName: "kubernetes.io/secret/9ce411af-8da3-4002-a75b-bc8ba4a64db8-secret-default-alertmanager-proxy-tls") pod "alertmanager-default-0" (UID: "9ce411af-8da3-4002-a75b-bc8ba4a64db8") : secret "default-alertmanager-proxy-tls" not found Oct 08 00:29:19 crc kubenswrapper[4967]: I1008 00:29:19.826853 4967 generic.go:334] "Generic (PLEG): container finished" podID="82ce08ad-02fd-4ae6-b3c5-c2c01ac6fdec" containerID="b67d9ba8ddaa2318500760e6ca0d7ad92cb9f715da966aa161799d7154559ca2" exitCode=0 Oct 08 00:29:19 crc kubenswrapper[4967]: I1008 00:29:19.826910 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-default-0" event={"ID":"82ce08ad-02fd-4ae6-b3c5-c2c01ac6fdec","Type":"ContainerDied","Data":"b67d9ba8ddaa2318500760e6ca0d7ad92cb9f715da966aa161799d7154559ca2"} Oct 08 00:29:20 crc kubenswrapper[4967]: I1008 00:29:20.227819 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-default-alertmanager-proxy-tls\" (UniqueName: \"kubernetes.io/secret/9ce411af-8da3-4002-a75b-bc8ba4a64db8-secret-default-alertmanager-proxy-tls\") pod \"alertmanager-default-0\" (UID: \"9ce411af-8da3-4002-a75b-bc8ba4a64db8\") " pod="service-telemetry/alertmanager-default-0" Oct 08 00:29:20 crc kubenswrapper[4967]: E1008 00:29:20.228276 4967 secret.go:188] Couldn't get secret service-telemetry/default-alertmanager-proxy-tls: secret "default-alertmanager-proxy-tls" not found Oct 08 00:29:20 crc kubenswrapper[4967]: E1008 00:29:20.228346 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9ce411af-8da3-4002-a75b-bc8ba4a64db8-secret-default-alertmanager-proxy-tls podName:9ce411af-8da3-4002-a75b-bc8ba4a64db8 nodeName:}" failed. No retries permitted until 2025-10-08 00:29:22.228328777 +0000 UTC m=+1360.878863426 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "secret-default-alertmanager-proxy-tls" (UniqueName: "kubernetes.io/secret/9ce411af-8da3-4002-a75b-bc8ba4a64db8-secret-default-alertmanager-proxy-tls") pod "alertmanager-default-0" (UID: "9ce411af-8da3-4002-a75b-bc8ba4a64db8") : secret "default-alertmanager-proxy-tls" not found Oct 08 00:29:21 crc kubenswrapper[4967]: I1008 00:29:21.843244 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-snmp-webhook-6856cfb745-6jgkj" event={"ID":"f3eb705e-460b-45f4-be46-8d811bda63c3","Type":"ContainerStarted","Data":"8b741b3c91d3f4897abb30ea82755e845dbebb09d88833f453aeb18a3549350f"} Oct 08 00:29:21 crc kubenswrapper[4967]: I1008 00:29:21.861862 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-snmp-webhook-6856cfb745-6jgkj" podStartSLOduration=1.682092415 podStartE2EDuration="7.861845854s" podCreationTimestamp="2025-10-08 00:29:14 +0000 UTC" firstStartedPulling="2025-10-08 00:29:15.17803418 +0000 UTC m=+1353.828568829" lastFinishedPulling="2025-10-08 00:29:21.357787619 +0000 UTC m=+1360.008322268" observedRunningTime="2025-10-08 00:29:21.857030603 +0000 UTC m=+1360.507565252" watchObservedRunningTime="2025-10-08 00:29:21.861845854 +0000 UTC m=+1360.512380493" Oct 08 00:29:22 crc kubenswrapper[4967]: I1008 00:29:22.255952 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-default-alertmanager-proxy-tls\" (UniqueName: \"kubernetes.io/secret/9ce411af-8da3-4002-a75b-bc8ba4a64db8-secret-default-alertmanager-proxy-tls\") pod \"alertmanager-default-0\" (UID: \"9ce411af-8da3-4002-a75b-bc8ba4a64db8\") " pod="service-telemetry/alertmanager-default-0" Oct 08 00:29:22 crc kubenswrapper[4967]: I1008 00:29:22.275637 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-default-alertmanager-proxy-tls\" (UniqueName: \"kubernetes.io/secret/9ce411af-8da3-4002-a75b-bc8ba4a64db8-secret-default-alertmanager-proxy-tls\") pod \"alertmanager-default-0\" (UID: \"9ce411af-8da3-4002-a75b-bc8ba4a64db8\") " pod="service-telemetry/alertmanager-default-0" Oct 08 00:29:22 crc kubenswrapper[4967]: I1008 00:29:22.419370 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/alertmanager-default-0" Oct 08 00:29:22 crc kubenswrapper[4967]: I1008 00:29:22.826461 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/alertmanager-default-0"] Oct 08 00:29:24 crc kubenswrapper[4967]: W1008 00:29:24.099974 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9ce411af_8da3_4002_a75b_bc8ba4a64db8.slice/crio-357dd91210fbed6c9c7531efa0558fd206b5c3cdfa20cc2523585e242bcbf7fd WatchSource:0}: Error finding container 357dd91210fbed6c9c7531efa0558fd206b5c3cdfa20cc2523585e242bcbf7fd: Status 404 returned error can't find the container with id 357dd91210fbed6c9c7531efa0558fd206b5c3cdfa20cc2523585e242bcbf7fd Oct 08 00:29:24 crc kubenswrapper[4967]: I1008 00:29:24.861967 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-default-0" event={"ID":"82ce08ad-02fd-4ae6-b3c5-c2c01ac6fdec","Type":"ContainerStarted","Data":"2dabbc18884ceb265f7a8694c373a8774a0ac7a0c4f85586980d04d15d1bdce9"} Oct 08 00:29:24 crc kubenswrapper[4967]: I1008 00:29:24.863979 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/alertmanager-default-0" event={"ID":"9ce411af-8da3-4002-a75b-bc8ba4a64db8","Type":"ContainerStarted","Data":"357dd91210fbed6c9c7531efa0558fd206b5c3cdfa20cc2523585e242bcbf7fd"} Oct 08 00:29:26 crc kubenswrapper[4967]: I1008 00:29:26.884044 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-default-0" event={"ID":"82ce08ad-02fd-4ae6-b3c5-c2c01ac6fdec","Type":"ContainerStarted","Data":"d946f0762a803e41c021bbe5ce3ee0467fc00da6fdcbaa890c03854394ec8e95"} Oct 08 00:29:26 crc kubenswrapper[4967]: I1008 00:29:26.886915 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/alertmanager-default-0" event={"ID":"9ce411af-8da3-4002-a75b-bc8ba4a64db8","Type":"ContainerStarted","Data":"e0e445694c7994e09b42aed09b690eac78b0302c98e618e539d571a1e151d3af"} Oct 08 00:29:30 crc kubenswrapper[4967]: I1008 00:29:30.695721 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-qcv8n"] Oct 08 00:29:30 crc kubenswrapper[4967]: I1008 00:29:30.697479 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-qcv8n" Oct 08 00:29:30 crc kubenswrapper[4967]: I1008 00:29:30.700813 4967 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"smart-gateway-session-secret" Oct 08 00:29:30 crc kubenswrapper[4967]: I1008 00:29:30.701113 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"default-cloud1-coll-meter-sg-core-configmap" Oct 08 00:29:30 crc kubenswrapper[4967]: I1008 00:29:30.701137 4967 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"smart-gateway-dockercfg-6wtnb" Oct 08 00:29:30 crc kubenswrapper[4967]: I1008 00:29:30.701323 4967 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-cloud1-coll-meter-proxy-tls" Oct 08 00:29:30 crc kubenswrapper[4967]: I1008 00:29:30.724949 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-qcv8n"] Oct 08 00:29:30 crc kubenswrapper[4967]: I1008 00:29:30.784603 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jbp8z\" (UniqueName: \"kubernetes.io/projected/ddba1803-397c-4daf-bc56-4857967229d1-kube-api-access-jbp8z\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-qcv8n\" (UID: \"ddba1803-397c-4daf-bc56-4857967229d1\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-qcv8n" Oct 08 00:29:30 crc kubenswrapper[4967]: I1008 00:29:30.784845 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/ddba1803-397c-4daf-bc56-4857967229d1-sg-core-config\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-qcv8n\" (UID: \"ddba1803-397c-4daf-bc56-4857967229d1\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-qcv8n" Oct 08 00:29:30 crc kubenswrapper[4967]: I1008 00:29:30.784892 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-cloud1-coll-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/ddba1803-397c-4daf-bc56-4857967229d1-default-cloud1-coll-meter-proxy-tls\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-qcv8n\" (UID: \"ddba1803-397c-4daf-bc56-4857967229d1\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-qcv8n" Oct 08 00:29:30 crc kubenswrapper[4967]: I1008 00:29:30.784928 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/ddba1803-397c-4daf-bc56-4857967229d1-socket-dir\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-qcv8n\" (UID: \"ddba1803-397c-4daf-bc56-4857967229d1\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-qcv8n" Oct 08 00:29:30 crc kubenswrapper[4967]: I1008 00:29:30.784996 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/ddba1803-397c-4daf-bc56-4857967229d1-session-secret\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-qcv8n\" (UID: \"ddba1803-397c-4daf-bc56-4857967229d1\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-qcv8n" Oct 08 00:29:30 crc kubenswrapper[4967]: I1008 00:29:30.886686 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/ddba1803-397c-4daf-bc56-4857967229d1-socket-dir\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-qcv8n\" (UID: \"ddba1803-397c-4daf-bc56-4857967229d1\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-qcv8n" Oct 08 00:29:30 crc kubenswrapper[4967]: I1008 00:29:30.886742 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/ddba1803-397c-4daf-bc56-4857967229d1-session-secret\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-qcv8n\" (UID: \"ddba1803-397c-4daf-bc56-4857967229d1\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-qcv8n" Oct 08 00:29:30 crc kubenswrapper[4967]: I1008 00:29:30.886770 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jbp8z\" (UniqueName: \"kubernetes.io/projected/ddba1803-397c-4daf-bc56-4857967229d1-kube-api-access-jbp8z\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-qcv8n\" (UID: \"ddba1803-397c-4daf-bc56-4857967229d1\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-qcv8n" Oct 08 00:29:30 crc kubenswrapper[4967]: I1008 00:29:30.886836 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/ddba1803-397c-4daf-bc56-4857967229d1-sg-core-config\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-qcv8n\" (UID: \"ddba1803-397c-4daf-bc56-4857967229d1\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-qcv8n" Oct 08 00:29:30 crc kubenswrapper[4967]: I1008 00:29:30.886865 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-cloud1-coll-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/ddba1803-397c-4daf-bc56-4857967229d1-default-cloud1-coll-meter-proxy-tls\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-qcv8n\" (UID: \"ddba1803-397c-4daf-bc56-4857967229d1\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-qcv8n" Oct 08 00:29:30 crc kubenswrapper[4967]: E1008 00:29:30.886976 4967 secret.go:188] Couldn't get secret service-telemetry/default-cloud1-coll-meter-proxy-tls: secret "default-cloud1-coll-meter-proxy-tls" not found Oct 08 00:29:30 crc kubenswrapper[4967]: E1008 00:29:30.887025 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ddba1803-397c-4daf-bc56-4857967229d1-default-cloud1-coll-meter-proxy-tls podName:ddba1803-397c-4daf-bc56-4857967229d1 nodeName:}" failed. No retries permitted until 2025-10-08 00:29:31.387009634 +0000 UTC m=+1370.037544283 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "default-cloud1-coll-meter-proxy-tls" (UniqueName: "kubernetes.io/secret/ddba1803-397c-4daf-bc56-4857967229d1-default-cloud1-coll-meter-proxy-tls") pod "default-cloud1-coll-meter-smartgateway-7cd87f9766-qcv8n" (UID: "ddba1803-397c-4daf-bc56-4857967229d1") : secret "default-cloud1-coll-meter-proxy-tls" not found Oct 08 00:29:30 crc kubenswrapper[4967]: I1008 00:29:30.887312 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/ddba1803-397c-4daf-bc56-4857967229d1-socket-dir\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-qcv8n\" (UID: \"ddba1803-397c-4daf-bc56-4857967229d1\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-qcv8n" Oct 08 00:29:30 crc kubenswrapper[4967]: I1008 00:29:30.889498 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/ddba1803-397c-4daf-bc56-4857967229d1-sg-core-config\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-qcv8n\" (UID: \"ddba1803-397c-4daf-bc56-4857967229d1\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-qcv8n" Oct 08 00:29:30 crc kubenswrapper[4967]: I1008 00:29:30.896830 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/ddba1803-397c-4daf-bc56-4857967229d1-session-secret\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-qcv8n\" (UID: \"ddba1803-397c-4daf-bc56-4857967229d1\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-qcv8n" Oct 08 00:29:30 crc kubenswrapper[4967]: I1008 00:29:30.909323 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jbp8z\" (UniqueName: \"kubernetes.io/projected/ddba1803-397c-4daf-bc56-4857967229d1-kube-api-access-jbp8z\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-qcv8n\" (UID: \"ddba1803-397c-4daf-bc56-4857967229d1\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-qcv8n" Oct 08 00:29:31 crc kubenswrapper[4967]: I1008 00:29:31.393329 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-cloud1-coll-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/ddba1803-397c-4daf-bc56-4857967229d1-default-cloud1-coll-meter-proxy-tls\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-qcv8n\" (UID: \"ddba1803-397c-4daf-bc56-4857967229d1\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-qcv8n" Oct 08 00:29:31 crc kubenswrapper[4967]: E1008 00:29:31.393837 4967 secret.go:188] Couldn't get secret service-telemetry/default-cloud1-coll-meter-proxy-tls: secret "default-cloud1-coll-meter-proxy-tls" not found Oct 08 00:29:31 crc kubenswrapper[4967]: E1008 00:29:31.393890 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ddba1803-397c-4daf-bc56-4857967229d1-default-cloud1-coll-meter-proxy-tls podName:ddba1803-397c-4daf-bc56-4857967229d1 nodeName:}" failed. No retries permitted until 2025-10-08 00:29:32.3938761 +0000 UTC m=+1371.044410749 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "default-cloud1-coll-meter-proxy-tls" (UniqueName: "kubernetes.io/secret/ddba1803-397c-4daf-bc56-4857967229d1-default-cloud1-coll-meter-proxy-tls") pod "default-cloud1-coll-meter-smartgateway-7cd87f9766-qcv8n" (UID: "ddba1803-397c-4daf-bc56-4857967229d1") : secret "default-cloud1-coll-meter-proxy-tls" not found Oct 08 00:29:31 crc kubenswrapper[4967]: I1008 00:29:31.927291 4967 generic.go:334] "Generic (PLEG): container finished" podID="9ce411af-8da3-4002-a75b-bc8ba4a64db8" containerID="e0e445694c7994e09b42aed09b690eac78b0302c98e618e539d571a1e151d3af" exitCode=0 Oct 08 00:29:31 crc kubenswrapper[4967]: I1008 00:29:31.927381 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/alertmanager-default-0" event={"ID":"9ce411af-8da3-4002-a75b-bc8ba4a64db8","Type":"ContainerDied","Data":"e0e445694c7994e09b42aed09b690eac78b0302c98e618e539d571a1e151d3af"} Oct 08 00:29:32 crc kubenswrapper[4967]: I1008 00:29:32.413480 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-cloud1-coll-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/ddba1803-397c-4daf-bc56-4857967229d1-default-cloud1-coll-meter-proxy-tls\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-qcv8n\" (UID: \"ddba1803-397c-4daf-bc56-4857967229d1\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-qcv8n" Oct 08 00:29:32 crc kubenswrapper[4967]: I1008 00:29:32.423483 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-cloud1-coll-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/ddba1803-397c-4daf-bc56-4857967229d1-default-cloud1-coll-meter-proxy-tls\") pod \"default-cloud1-coll-meter-smartgateway-7cd87f9766-qcv8n\" (UID: \"ddba1803-397c-4daf-bc56-4857967229d1\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-qcv8n" Oct 08 00:29:32 crc kubenswrapper[4967]: I1008 00:29:32.521479 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-qcv8n" Oct 08 00:29:32 crc kubenswrapper[4967]: I1008 00:29:32.912168 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-dlql4"] Oct 08 00:29:32 crc kubenswrapper[4967]: I1008 00:29:32.913446 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-dlql4" Oct 08 00:29:32 crc kubenswrapper[4967]: I1008 00:29:32.916906 4967 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-cloud1-ceil-meter-proxy-tls" Oct 08 00:29:32 crc kubenswrapper[4967]: I1008 00:29:32.917028 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"default-cloud1-ceil-meter-sg-core-configmap" Oct 08 00:29:32 crc kubenswrapper[4967]: I1008 00:29:32.926691 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-dlql4"] Oct 08 00:29:33 crc kubenswrapper[4967]: I1008 00:29:33.022369 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s6p4m\" (UniqueName: \"kubernetes.io/projected/1f1c83bb-ff16-45ff-8fe5-e1b7e9d5166b-kube-api-access-s6p4m\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-dlql4\" (UID: \"1f1c83bb-ff16-45ff-8fe5-e1b7e9d5166b\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-dlql4" Oct 08 00:29:33 crc kubenswrapper[4967]: I1008 00:29:33.023157 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/1f1c83bb-ff16-45ff-8fe5-e1b7e9d5166b-sg-core-config\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-dlql4\" (UID: \"1f1c83bb-ff16-45ff-8fe5-e1b7e9d5166b\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-dlql4" Oct 08 00:29:33 crc kubenswrapper[4967]: I1008 00:29:33.023394 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/1f1c83bb-ff16-45ff-8fe5-e1b7e9d5166b-session-secret\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-dlql4\" (UID: \"1f1c83bb-ff16-45ff-8fe5-e1b7e9d5166b\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-dlql4" Oct 08 00:29:33 crc kubenswrapper[4967]: I1008 00:29:33.025522 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-cloud1-ceil-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/1f1c83bb-ff16-45ff-8fe5-e1b7e9d5166b-default-cloud1-ceil-meter-proxy-tls\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-dlql4\" (UID: \"1f1c83bb-ff16-45ff-8fe5-e1b7e9d5166b\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-dlql4" Oct 08 00:29:33 crc kubenswrapper[4967]: I1008 00:29:33.025723 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/1f1c83bb-ff16-45ff-8fe5-e1b7e9d5166b-socket-dir\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-dlql4\" (UID: \"1f1c83bb-ff16-45ff-8fe5-e1b7e9d5166b\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-dlql4" Oct 08 00:29:33 crc kubenswrapper[4967]: I1008 00:29:33.127474 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-cloud1-ceil-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/1f1c83bb-ff16-45ff-8fe5-e1b7e9d5166b-default-cloud1-ceil-meter-proxy-tls\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-dlql4\" (UID: \"1f1c83bb-ff16-45ff-8fe5-e1b7e9d5166b\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-dlql4" Oct 08 00:29:33 crc kubenswrapper[4967]: I1008 00:29:33.128023 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/1f1c83bb-ff16-45ff-8fe5-e1b7e9d5166b-socket-dir\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-dlql4\" (UID: \"1f1c83bb-ff16-45ff-8fe5-e1b7e9d5166b\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-dlql4" Oct 08 00:29:33 crc kubenswrapper[4967]: I1008 00:29:33.128767 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s6p4m\" (UniqueName: \"kubernetes.io/projected/1f1c83bb-ff16-45ff-8fe5-e1b7e9d5166b-kube-api-access-s6p4m\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-dlql4\" (UID: \"1f1c83bb-ff16-45ff-8fe5-e1b7e9d5166b\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-dlql4" Oct 08 00:29:33 crc kubenswrapper[4967]: I1008 00:29:33.130104 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/1f1c83bb-ff16-45ff-8fe5-e1b7e9d5166b-sg-core-config\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-dlql4\" (UID: \"1f1c83bb-ff16-45ff-8fe5-e1b7e9d5166b\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-dlql4" Oct 08 00:29:33 crc kubenswrapper[4967]: I1008 00:29:33.130809 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/1f1c83bb-ff16-45ff-8fe5-e1b7e9d5166b-session-secret\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-dlql4\" (UID: \"1f1c83bb-ff16-45ff-8fe5-e1b7e9d5166b\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-dlql4" Oct 08 00:29:33 crc kubenswrapper[4967]: I1008 00:29:33.128636 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/1f1c83bb-ff16-45ff-8fe5-e1b7e9d5166b-socket-dir\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-dlql4\" (UID: \"1f1c83bb-ff16-45ff-8fe5-e1b7e9d5166b\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-dlql4" Oct 08 00:29:33 crc kubenswrapper[4967]: I1008 00:29:33.131218 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/1f1c83bb-ff16-45ff-8fe5-e1b7e9d5166b-sg-core-config\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-dlql4\" (UID: \"1f1c83bb-ff16-45ff-8fe5-e1b7e9d5166b\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-dlql4" Oct 08 00:29:33 crc kubenswrapper[4967]: E1008 00:29:33.127974 4967 secret.go:188] Couldn't get secret service-telemetry/default-cloud1-ceil-meter-proxy-tls: secret "default-cloud1-ceil-meter-proxy-tls" not found Oct 08 00:29:33 crc kubenswrapper[4967]: E1008 00:29:33.131519 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1f1c83bb-ff16-45ff-8fe5-e1b7e9d5166b-default-cloud1-ceil-meter-proxy-tls podName:1f1c83bb-ff16-45ff-8fe5-e1b7e9d5166b nodeName:}" failed. No retries permitted until 2025-10-08 00:29:33.631493986 +0000 UTC m=+1372.282028635 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "default-cloud1-ceil-meter-proxy-tls" (UniqueName: "kubernetes.io/secret/1f1c83bb-ff16-45ff-8fe5-e1b7e9d5166b-default-cloud1-ceil-meter-proxy-tls") pod "default-cloud1-ceil-meter-smartgateway-57948895dc-dlql4" (UID: "1f1c83bb-ff16-45ff-8fe5-e1b7e9d5166b") : secret "default-cloud1-ceil-meter-proxy-tls" not found Oct 08 00:29:33 crc kubenswrapper[4967]: I1008 00:29:33.133761 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/1f1c83bb-ff16-45ff-8fe5-e1b7e9d5166b-session-secret\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-dlql4\" (UID: \"1f1c83bb-ff16-45ff-8fe5-e1b7e9d5166b\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-dlql4" Oct 08 00:29:33 crc kubenswrapper[4967]: I1008 00:29:33.148215 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s6p4m\" (UniqueName: \"kubernetes.io/projected/1f1c83bb-ff16-45ff-8fe5-e1b7e9d5166b-kube-api-access-s6p4m\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-dlql4\" (UID: \"1f1c83bb-ff16-45ff-8fe5-e1b7e9d5166b\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-dlql4" Oct 08 00:29:33 crc kubenswrapper[4967]: I1008 00:29:33.637418 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-cloud1-ceil-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/1f1c83bb-ff16-45ff-8fe5-e1b7e9d5166b-default-cloud1-ceil-meter-proxy-tls\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-dlql4\" (UID: \"1f1c83bb-ff16-45ff-8fe5-e1b7e9d5166b\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-dlql4" Oct 08 00:29:33 crc kubenswrapper[4967]: E1008 00:29:33.637615 4967 secret.go:188] Couldn't get secret service-telemetry/default-cloud1-ceil-meter-proxy-tls: secret "default-cloud1-ceil-meter-proxy-tls" not found Oct 08 00:29:33 crc kubenswrapper[4967]: E1008 00:29:33.637732 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1f1c83bb-ff16-45ff-8fe5-e1b7e9d5166b-default-cloud1-ceil-meter-proxy-tls podName:1f1c83bb-ff16-45ff-8fe5-e1b7e9d5166b nodeName:}" failed. No retries permitted until 2025-10-08 00:29:34.637710855 +0000 UTC m=+1373.288245524 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "default-cloud1-ceil-meter-proxy-tls" (UniqueName: "kubernetes.io/secret/1f1c83bb-ff16-45ff-8fe5-e1b7e9d5166b-default-cloud1-ceil-meter-proxy-tls") pod "default-cloud1-ceil-meter-smartgateway-57948895dc-dlql4" (UID: "1f1c83bb-ff16-45ff-8fe5-e1b7e9d5166b") : secret "default-cloud1-ceil-meter-proxy-tls" not found Oct 08 00:29:34 crc kubenswrapper[4967]: I1008 00:29:34.651907 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-cloud1-ceil-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/1f1c83bb-ff16-45ff-8fe5-e1b7e9d5166b-default-cloud1-ceil-meter-proxy-tls\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-dlql4\" (UID: \"1f1c83bb-ff16-45ff-8fe5-e1b7e9d5166b\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-dlql4" Oct 08 00:29:34 crc kubenswrapper[4967]: I1008 00:29:34.675490 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-cloud1-ceil-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/1f1c83bb-ff16-45ff-8fe5-e1b7e9d5166b-default-cloud1-ceil-meter-proxy-tls\") pod \"default-cloud1-ceil-meter-smartgateway-57948895dc-dlql4\" (UID: \"1f1c83bb-ff16-45ff-8fe5-e1b7e9d5166b\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-dlql4" Oct 08 00:29:34 crc kubenswrapper[4967]: I1008 00:29:34.734148 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-dlql4" Oct 08 00:29:37 crc kubenswrapper[4967]: I1008 00:29:37.259451 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-rwnt8"] Oct 08 00:29:37 crc kubenswrapper[4967]: I1008 00:29:37.261550 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-rwnt8" Oct 08 00:29:37 crc kubenswrapper[4967]: I1008 00:29:37.264866 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"default-cloud1-sens-meter-sg-core-configmap" Oct 08 00:29:37 crc kubenswrapper[4967]: I1008 00:29:37.266616 4967 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-cloud1-sens-meter-proxy-tls" Oct 08 00:29:37 crc kubenswrapper[4967]: I1008 00:29:37.267826 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-rwnt8"] Oct 08 00:29:37 crc kubenswrapper[4967]: I1008 00:29:37.396125 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/a9da825c-48d2-477d-b8c5-1ac878239c37-session-secret\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-rwnt8\" (UID: \"a9da825c-48d2-477d-b8c5-1ac878239c37\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-rwnt8" Oct 08 00:29:37 crc kubenswrapper[4967]: I1008 00:29:37.396258 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/a9da825c-48d2-477d-b8c5-1ac878239c37-socket-dir\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-rwnt8\" (UID: \"a9da825c-48d2-477d-b8c5-1ac878239c37\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-rwnt8" Oct 08 00:29:37 crc kubenswrapper[4967]: I1008 00:29:37.396295 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-cloud1-sens-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/a9da825c-48d2-477d-b8c5-1ac878239c37-default-cloud1-sens-meter-proxy-tls\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-rwnt8\" (UID: \"a9da825c-48d2-477d-b8c5-1ac878239c37\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-rwnt8" Oct 08 00:29:37 crc kubenswrapper[4967]: I1008 00:29:37.396368 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/a9da825c-48d2-477d-b8c5-1ac878239c37-sg-core-config\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-rwnt8\" (UID: \"a9da825c-48d2-477d-b8c5-1ac878239c37\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-rwnt8" Oct 08 00:29:37 crc kubenswrapper[4967]: I1008 00:29:37.396393 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tkkrk\" (UniqueName: \"kubernetes.io/projected/a9da825c-48d2-477d-b8c5-1ac878239c37-kube-api-access-tkkrk\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-rwnt8\" (UID: \"a9da825c-48d2-477d-b8c5-1ac878239c37\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-rwnt8" Oct 08 00:29:37 crc kubenswrapper[4967]: I1008 00:29:37.498372 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/a9da825c-48d2-477d-b8c5-1ac878239c37-socket-dir\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-rwnt8\" (UID: \"a9da825c-48d2-477d-b8c5-1ac878239c37\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-rwnt8" Oct 08 00:29:37 crc kubenswrapper[4967]: I1008 00:29:37.498418 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-cloud1-sens-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/a9da825c-48d2-477d-b8c5-1ac878239c37-default-cloud1-sens-meter-proxy-tls\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-rwnt8\" (UID: \"a9da825c-48d2-477d-b8c5-1ac878239c37\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-rwnt8" Oct 08 00:29:37 crc kubenswrapper[4967]: I1008 00:29:37.498448 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/a9da825c-48d2-477d-b8c5-1ac878239c37-sg-core-config\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-rwnt8\" (UID: \"a9da825c-48d2-477d-b8c5-1ac878239c37\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-rwnt8" Oct 08 00:29:37 crc kubenswrapper[4967]: I1008 00:29:37.498473 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tkkrk\" (UniqueName: \"kubernetes.io/projected/a9da825c-48d2-477d-b8c5-1ac878239c37-kube-api-access-tkkrk\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-rwnt8\" (UID: \"a9da825c-48d2-477d-b8c5-1ac878239c37\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-rwnt8" Oct 08 00:29:37 crc kubenswrapper[4967]: I1008 00:29:37.498544 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/a9da825c-48d2-477d-b8c5-1ac878239c37-session-secret\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-rwnt8\" (UID: \"a9da825c-48d2-477d-b8c5-1ac878239c37\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-rwnt8" Oct 08 00:29:37 crc kubenswrapper[4967]: I1008 00:29:37.499080 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/a9da825c-48d2-477d-b8c5-1ac878239c37-socket-dir\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-rwnt8\" (UID: \"a9da825c-48d2-477d-b8c5-1ac878239c37\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-rwnt8" Oct 08 00:29:37 crc kubenswrapper[4967]: E1008 00:29:37.499236 4967 secret.go:188] Couldn't get secret service-telemetry/default-cloud1-sens-meter-proxy-tls: secret "default-cloud1-sens-meter-proxy-tls" not found Oct 08 00:29:37 crc kubenswrapper[4967]: E1008 00:29:37.499324 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a9da825c-48d2-477d-b8c5-1ac878239c37-default-cloud1-sens-meter-proxy-tls podName:a9da825c-48d2-477d-b8c5-1ac878239c37 nodeName:}" failed. No retries permitted until 2025-10-08 00:29:37.999303631 +0000 UTC m=+1376.649838280 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "default-cloud1-sens-meter-proxy-tls" (UniqueName: "kubernetes.io/secret/a9da825c-48d2-477d-b8c5-1ac878239c37-default-cloud1-sens-meter-proxy-tls") pod "default-cloud1-sens-meter-smartgateway-5759b4d97-rwnt8" (UID: "a9da825c-48d2-477d-b8c5-1ac878239c37") : secret "default-cloud1-sens-meter-proxy-tls" not found Oct 08 00:29:37 crc kubenswrapper[4967]: I1008 00:29:37.500235 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/a9da825c-48d2-477d-b8c5-1ac878239c37-sg-core-config\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-rwnt8\" (UID: \"a9da825c-48d2-477d-b8c5-1ac878239c37\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-rwnt8" Oct 08 00:29:37 crc kubenswrapper[4967]: I1008 00:29:37.503957 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/a9da825c-48d2-477d-b8c5-1ac878239c37-session-secret\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-rwnt8\" (UID: \"a9da825c-48d2-477d-b8c5-1ac878239c37\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-rwnt8" Oct 08 00:29:37 crc kubenswrapper[4967]: I1008 00:29:37.516858 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tkkrk\" (UniqueName: \"kubernetes.io/projected/a9da825c-48d2-477d-b8c5-1ac878239c37-kube-api-access-tkkrk\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-rwnt8\" (UID: \"a9da825c-48d2-477d-b8c5-1ac878239c37\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-rwnt8" Oct 08 00:29:38 crc kubenswrapper[4967]: I1008 00:29:38.015922 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-cloud1-sens-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/a9da825c-48d2-477d-b8c5-1ac878239c37-default-cloud1-sens-meter-proxy-tls\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-rwnt8\" (UID: \"a9da825c-48d2-477d-b8c5-1ac878239c37\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-rwnt8" Oct 08 00:29:38 crc kubenswrapper[4967]: E1008 00:29:38.016074 4967 secret.go:188] Couldn't get secret service-telemetry/default-cloud1-sens-meter-proxy-tls: secret "default-cloud1-sens-meter-proxy-tls" not found Oct 08 00:29:38 crc kubenswrapper[4967]: E1008 00:29:38.016158 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a9da825c-48d2-477d-b8c5-1ac878239c37-default-cloud1-sens-meter-proxy-tls podName:a9da825c-48d2-477d-b8c5-1ac878239c37 nodeName:}" failed. No retries permitted until 2025-10-08 00:29:39.016138837 +0000 UTC m=+1377.666673496 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "default-cloud1-sens-meter-proxy-tls" (UniqueName: "kubernetes.io/secret/a9da825c-48d2-477d-b8c5-1ac878239c37-default-cloud1-sens-meter-proxy-tls") pod "default-cloud1-sens-meter-smartgateway-5759b4d97-rwnt8" (UID: "a9da825c-48d2-477d-b8c5-1ac878239c37") : secret "default-cloud1-sens-meter-proxy-tls" not found Oct 08 00:29:38 crc kubenswrapper[4967]: I1008 00:29:38.671610 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-qcv8n"] Oct 08 00:29:38 crc kubenswrapper[4967]: I1008 00:29:38.826845 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-dlql4"] Oct 08 00:29:38 crc kubenswrapper[4967]: W1008 00:29:38.854972 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1f1c83bb_ff16_45ff_8fe5_e1b7e9d5166b.slice/crio-7b472bfed4ca20f0ca5cccb64132aa7a6a2c9fc5e91f0b42eacfaea2bdd72009 WatchSource:0}: Error finding container 7b472bfed4ca20f0ca5cccb64132aa7a6a2c9fc5e91f0b42eacfaea2bdd72009: Status 404 returned error can't find the container with id 7b472bfed4ca20f0ca5cccb64132aa7a6a2c9fc5e91f0b42eacfaea2bdd72009 Oct 08 00:29:38 crc kubenswrapper[4967]: I1008 00:29:38.969248 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-qcv8n" event={"ID":"ddba1803-397c-4daf-bc56-4857967229d1","Type":"ContainerStarted","Data":"3178f706c46e067e27b620fe08ebb2580c6df6ccc3992cbf0a3efc5caddbb946"} Oct 08 00:29:38 crc kubenswrapper[4967]: I1008 00:29:38.973003 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-default-0" event={"ID":"82ce08ad-02fd-4ae6-b3c5-c2c01ac6fdec","Type":"ContainerStarted","Data":"c1f9ff415cb43670f56c92f7a16bad63b13bcb2f669fc863714ec9d943f041b9"} Oct 08 00:29:38 crc kubenswrapper[4967]: I1008 00:29:38.975326 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-dlql4" event={"ID":"1f1c83bb-ff16-45ff-8fe5-e1b7e9d5166b","Type":"ContainerStarted","Data":"7b472bfed4ca20f0ca5cccb64132aa7a6a2c9fc5e91f0b42eacfaea2bdd72009"} Oct 08 00:29:39 crc kubenswrapper[4967]: I1008 00:29:39.033385 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-cloud1-sens-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/a9da825c-48d2-477d-b8c5-1ac878239c37-default-cloud1-sens-meter-proxy-tls\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-rwnt8\" (UID: \"a9da825c-48d2-477d-b8c5-1ac878239c37\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-rwnt8" Oct 08 00:29:39 crc kubenswrapper[4967]: I1008 00:29:39.040229 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-cloud1-sens-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/a9da825c-48d2-477d-b8c5-1ac878239c37-default-cloud1-sens-meter-proxy-tls\") pod \"default-cloud1-sens-meter-smartgateway-5759b4d97-rwnt8\" (UID: \"a9da825c-48d2-477d-b8c5-1ac878239c37\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-rwnt8" Oct 08 00:29:39 crc kubenswrapper[4967]: I1008 00:29:39.109419 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-rwnt8" Oct 08 00:29:40 crc kubenswrapper[4967]: I1008 00:29:40.450030 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/prometheus-default-0" podStartSLOduration=6.087713819 podStartE2EDuration="37.450013241s" podCreationTimestamp="2025-10-08 00:29:03 +0000 UTC" firstStartedPulling="2025-10-08 00:29:07.058122482 +0000 UTC m=+1345.708657151" lastFinishedPulling="2025-10-08 00:29:38.420421924 +0000 UTC m=+1377.070956573" observedRunningTime="2025-10-08 00:29:38.99651419 +0000 UTC m=+1377.647048859" watchObservedRunningTime="2025-10-08 00:29:40.450013241 +0000 UTC m=+1379.100547890" Oct 08 00:29:40 crc kubenswrapper[4967]: I1008 00:29:40.452824 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-rwnt8"] Oct 08 00:29:40 crc kubenswrapper[4967]: I1008 00:29:40.991141 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/alertmanager-default-0" event={"ID":"9ce411af-8da3-4002-a75b-bc8ba4a64db8","Type":"ContainerStarted","Data":"a5ce0126d2a069a2a73606cf8ff87ebad0dc3cf3057c379eb0be9f4c2b8a2721"} Oct 08 00:29:40 crc kubenswrapper[4967]: I1008 00:29:40.992643 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-rwnt8" event={"ID":"a9da825c-48d2-477d-b8c5-1ac878239c37","Type":"ContainerStarted","Data":"9e2ca9168ffc0a6883299a8d30f0ec89b22e61be02fe536e9770594cb12179cd"} Oct 08 00:29:40 crc kubenswrapper[4967]: I1008 00:29:40.994348 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-qcv8n" event={"ID":"ddba1803-397c-4daf-bc56-4857967229d1","Type":"ContainerStarted","Data":"f6586ce7a964efd2b140aa8e54dddc87af488f192dcc47f7d34749fdc44157d5"} Oct 08 00:29:40 crc kubenswrapper[4967]: I1008 00:29:40.995445 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-dlql4" event={"ID":"1f1c83bb-ff16-45ff-8fe5-e1b7e9d5166b","Type":"ContainerStarted","Data":"7e24fa230e3ce8d03f327be71ae6e8a5f5078ae4e95fc98cbcf196e40c1b8909"} Oct 08 00:29:41 crc kubenswrapper[4967]: I1008 00:29:41.605936 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="service-telemetry/prometheus-default-0" Oct 08 00:29:42 crc kubenswrapper[4967]: I1008 00:29:42.003357 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-rwnt8" event={"ID":"a9da825c-48d2-477d-b8c5-1ac878239c37","Type":"ContainerStarted","Data":"a5b2ea50ad497df15eaff2bb5bbbf806c8abd3a2763dde2647ee933d98467f6e"} Oct 08 00:29:43 crc kubenswrapper[4967]: I1008 00:29:43.015061 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/alertmanager-default-0" event={"ID":"9ce411af-8da3-4002-a75b-bc8ba4a64db8","Type":"ContainerStarted","Data":"ce4f79a4fe539c88e2994d2b472f4ecc0628fde63d2bbf421a46337ca7680700"} Oct 08 00:29:44 crc kubenswrapper[4967]: I1008 00:29:44.097491 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/default-cloud1-coll-event-smartgateway-6f7464cc9d-bjhng"] Oct 08 00:29:44 crc kubenswrapper[4967]: I1008 00:29:44.099299 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-coll-event-smartgateway-6f7464cc9d-bjhng" Oct 08 00:29:44 crc kubenswrapper[4967]: I1008 00:29:44.105220 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"default-cloud1-coll-event-sg-core-configmap" Oct 08 00:29:44 crc kubenswrapper[4967]: I1008 00:29:44.105585 4967 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-cert" Oct 08 00:29:44 crc kubenswrapper[4967]: I1008 00:29:44.134011 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-coll-event-smartgateway-6f7464cc9d-bjhng"] Oct 08 00:29:44 crc kubenswrapper[4967]: I1008 00:29:44.207984 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q4s7l\" (UniqueName: \"kubernetes.io/projected/84dadc03-b18c-490c-adb2-9d3f605865d1-kube-api-access-q4s7l\") pod \"default-cloud1-coll-event-smartgateway-6f7464cc9d-bjhng\" (UID: \"84dadc03-b18c-490c-adb2-9d3f605865d1\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-6f7464cc9d-bjhng" Oct 08 00:29:44 crc kubenswrapper[4967]: I1008 00:29:44.208046 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/84dadc03-b18c-490c-adb2-9d3f605865d1-sg-core-config\") pod \"default-cloud1-coll-event-smartgateway-6f7464cc9d-bjhng\" (UID: \"84dadc03-b18c-490c-adb2-9d3f605865d1\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-6f7464cc9d-bjhng" Oct 08 00:29:44 crc kubenswrapper[4967]: I1008 00:29:44.208074 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/84dadc03-b18c-490c-adb2-9d3f605865d1-socket-dir\") pod \"default-cloud1-coll-event-smartgateway-6f7464cc9d-bjhng\" (UID: \"84dadc03-b18c-490c-adb2-9d3f605865d1\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-6f7464cc9d-bjhng" Oct 08 00:29:44 crc kubenswrapper[4967]: I1008 00:29:44.208127 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-certs\" (UniqueName: \"kubernetes.io/secret/84dadc03-b18c-490c-adb2-9d3f605865d1-elastic-certs\") pod \"default-cloud1-coll-event-smartgateway-6f7464cc9d-bjhng\" (UID: \"84dadc03-b18c-490c-adb2-9d3f605865d1\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-6f7464cc9d-bjhng" Oct 08 00:29:44 crc kubenswrapper[4967]: I1008 00:29:44.309659 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q4s7l\" (UniqueName: \"kubernetes.io/projected/84dadc03-b18c-490c-adb2-9d3f605865d1-kube-api-access-q4s7l\") pod \"default-cloud1-coll-event-smartgateway-6f7464cc9d-bjhng\" (UID: \"84dadc03-b18c-490c-adb2-9d3f605865d1\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-6f7464cc9d-bjhng" Oct 08 00:29:44 crc kubenswrapper[4967]: I1008 00:29:44.309757 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/84dadc03-b18c-490c-adb2-9d3f605865d1-sg-core-config\") pod \"default-cloud1-coll-event-smartgateway-6f7464cc9d-bjhng\" (UID: \"84dadc03-b18c-490c-adb2-9d3f605865d1\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-6f7464cc9d-bjhng" Oct 08 00:29:44 crc kubenswrapper[4967]: I1008 00:29:44.309787 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/84dadc03-b18c-490c-adb2-9d3f605865d1-socket-dir\") pod \"default-cloud1-coll-event-smartgateway-6f7464cc9d-bjhng\" (UID: \"84dadc03-b18c-490c-adb2-9d3f605865d1\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-6f7464cc9d-bjhng" Oct 08 00:29:44 crc kubenswrapper[4967]: I1008 00:29:44.309845 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-certs\" (UniqueName: \"kubernetes.io/secret/84dadc03-b18c-490c-adb2-9d3f605865d1-elastic-certs\") pod \"default-cloud1-coll-event-smartgateway-6f7464cc9d-bjhng\" (UID: \"84dadc03-b18c-490c-adb2-9d3f605865d1\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-6f7464cc9d-bjhng" Oct 08 00:29:44 crc kubenswrapper[4967]: I1008 00:29:44.310987 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/84dadc03-b18c-490c-adb2-9d3f605865d1-socket-dir\") pod \"default-cloud1-coll-event-smartgateway-6f7464cc9d-bjhng\" (UID: \"84dadc03-b18c-490c-adb2-9d3f605865d1\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-6f7464cc9d-bjhng" Oct 08 00:29:44 crc kubenswrapper[4967]: I1008 00:29:44.314167 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/84dadc03-b18c-490c-adb2-9d3f605865d1-sg-core-config\") pod \"default-cloud1-coll-event-smartgateway-6f7464cc9d-bjhng\" (UID: \"84dadc03-b18c-490c-adb2-9d3f605865d1\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-6f7464cc9d-bjhng" Oct 08 00:29:44 crc kubenswrapper[4967]: I1008 00:29:44.315178 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-certs\" (UniqueName: \"kubernetes.io/secret/84dadc03-b18c-490c-adb2-9d3f605865d1-elastic-certs\") pod \"default-cloud1-coll-event-smartgateway-6f7464cc9d-bjhng\" (UID: \"84dadc03-b18c-490c-adb2-9d3f605865d1\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-6f7464cc9d-bjhng" Oct 08 00:29:44 crc kubenswrapper[4967]: I1008 00:29:44.327192 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q4s7l\" (UniqueName: \"kubernetes.io/projected/84dadc03-b18c-490c-adb2-9d3f605865d1-kube-api-access-q4s7l\") pod \"default-cloud1-coll-event-smartgateway-6f7464cc9d-bjhng\" (UID: \"84dadc03-b18c-490c-adb2-9d3f605865d1\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-6f7464cc9d-bjhng" Oct 08 00:29:44 crc kubenswrapper[4967]: I1008 00:29:44.417420 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-coll-event-smartgateway-6f7464cc9d-bjhng" Oct 08 00:29:44 crc kubenswrapper[4967]: I1008 00:29:44.556173 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/default-cloud1-ceil-event-smartgateway-5d47c9955c-t24dp"] Oct 08 00:29:44 crc kubenswrapper[4967]: I1008 00:29:44.558890 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-5d47c9955c-t24dp" Oct 08 00:29:44 crc kubenswrapper[4967]: I1008 00:29:44.563194 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"default-cloud1-ceil-event-sg-core-configmap" Oct 08 00:29:44 crc kubenswrapper[4967]: I1008 00:29:44.582922 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-ceil-event-smartgateway-5d47c9955c-t24dp"] Oct 08 00:29:44 crc kubenswrapper[4967]: I1008 00:29:44.718057 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-certs\" (UniqueName: \"kubernetes.io/secret/b3e181c2-79f9-47be-b4fd-f9f05599395f-elastic-certs\") pod \"default-cloud1-ceil-event-smartgateway-5d47c9955c-t24dp\" (UID: \"b3e181c2-79f9-47be-b4fd-f9f05599395f\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-5d47c9955c-t24dp" Oct 08 00:29:44 crc kubenswrapper[4967]: I1008 00:29:44.718186 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8j6ns\" (UniqueName: \"kubernetes.io/projected/b3e181c2-79f9-47be-b4fd-f9f05599395f-kube-api-access-8j6ns\") pod \"default-cloud1-ceil-event-smartgateway-5d47c9955c-t24dp\" (UID: \"b3e181c2-79f9-47be-b4fd-f9f05599395f\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-5d47c9955c-t24dp" Oct 08 00:29:44 crc kubenswrapper[4967]: I1008 00:29:44.718369 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/b3e181c2-79f9-47be-b4fd-f9f05599395f-sg-core-config\") pod \"default-cloud1-ceil-event-smartgateway-5d47c9955c-t24dp\" (UID: \"b3e181c2-79f9-47be-b4fd-f9f05599395f\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-5d47c9955c-t24dp" Oct 08 00:29:44 crc kubenswrapper[4967]: I1008 00:29:44.718423 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/b3e181c2-79f9-47be-b4fd-f9f05599395f-socket-dir\") pod \"default-cloud1-ceil-event-smartgateway-5d47c9955c-t24dp\" (UID: \"b3e181c2-79f9-47be-b4fd-f9f05599395f\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-5d47c9955c-t24dp" Oct 08 00:29:44 crc kubenswrapper[4967]: I1008 00:29:44.820219 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/b3e181c2-79f9-47be-b4fd-f9f05599395f-sg-core-config\") pod \"default-cloud1-ceil-event-smartgateway-5d47c9955c-t24dp\" (UID: \"b3e181c2-79f9-47be-b4fd-f9f05599395f\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-5d47c9955c-t24dp" Oct 08 00:29:44 crc kubenswrapper[4967]: I1008 00:29:44.820261 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/b3e181c2-79f9-47be-b4fd-f9f05599395f-socket-dir\") pod \"default-cloud1-ceil-event-smartgateway-5d47c9955c-t24dp\" (UID: \"b3e181c2-79f9-47be-b4fd-f9f05599395f\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-5d47c9955c-t24dp" Oct 08 00:29:44 crc kubenswrapper[4967]: I1008 00:29:44.820283 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-certs\" (UniqueName: \"kubernetes.io/secret/b3e181c2-79f9-47be-b4fd-f9f05599395f-elastic-certs\") pod \"default-cloud1-ceil-event-smartgateway-5d47c9955c-t24dp\" (UID: \"b3e181c2-79f9-47be-b4fd-f9f05599395f\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-5d47c9955c-t24dp" Oct 08 00:29:44 crc kubenswrapper[4967]: I1008 00:29:44.820333 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8j6ns\" (UniqueName: \"kubernetes.io/projected/b3e181c2-79f9-47be-b4fd-f9f05599395f-kube-api-access-8j6ns\") pod \"default-cloud1-ceil-event-smartgateway-5d47c9955c-t24dp\" (UID: \"b3e181c2-79f9-47be-b4fd-f9f05599395f\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-5d47c9955c-t24dp" Oct 08 00:29:44 crc kubenswrapper[4967]: I1008 00:29:44.820880 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/b3e181c2-79f9-47be-b4fd-f9f05599395f-socket-dir\") pod \"default-cloud1-ceil-event-smartgateway-5d47c9955c-t24dp\" (UID: \"b3e181c2-79f9-47be-b4fd-f9f05599395f\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-5d47c9955c-t24dp" Oct 08 00:29:44 crc kubenswrapper[4967]: I1008 00:29:44.821212 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/b3e181c2-79f9-47be-b4fd-f9f05599395f-sg-core-config\") pod \"default-cloud1-ceil-event-smartgateway-5d47c9955c-t24dp\" (UID: \"b3e181c2-79f9-47be-b4fd-f9f05599395f\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-5d47c9955c-t24dp" Oct 08 00:29:44 crc kubenswrapper[4967]: I1008 00:29:44.824230 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-certs\" (UniqueName: \"kubernetes.io/secret/b3e181c2-79f9-47be-b4fd-f9f05599395f-elastic-certs\") pod \"default-cloud1-ceil-event-smartgateway-5d47c9955c-t24dp\" (UID: \"b3e181c2-79f9-47be-b4fd-f9f05599395f\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-5d47c9955c-t24dp" Oct 08 00:29:44 crc kubenswrapper[4967]: I1008 00:29:44.853643 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8j6ns\" (UniqueName: \"kubernetes.io/projected/b3e181c2-79f9-47be-b4fd-f9f05599395f-kube-api-access-8j6ns\") pod \"default-cloud1-ceil-event-smartgateway-5d47c9955c-t24dp\" (UID: \"b3e181c2-79f9-47be-b4fd-f9f05599395f\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-5d47c9955c-t24dp" Oct 08 00:29:44 crc kubenswrapper[4967]: I1008 00:29:44.897520 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-5d47c9955c-t24dp" Oct 08 00:29:46 crc kubenswrapper[4967]: I1008 00:29:46.040078 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/alertmanager-default-0" event={"ID":"9ce411af-8da3-4002-a75b-bc8ba4a64db8","Type":"ContainerStarted","Data":"8c4a4933f68800e29db56d85b03062d09a6850714fb15a603925a3415fd2edec"} Oct 08 00:29:46 crc kubenswrapper[4967]: I1008 00:29:46.055996 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-rwnt8" event={"ID":"a9da825c-48d2-477d-b8c5-1ac878239c37","Type":"ContainerStarted","Data":"0526ba518406905f366079730b0681ac33f7b4d3625854c26e781db9e85f01d1"} Oct 08 00:29:46 crc kubenswrapper[4967]: I1008 00:29:46.068746 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-qcv8n" event={"ID":"ddba1803-397c-4daf-bc56-4857967229d1","Type":"ContainerStarted","Data":"3a11ae15580d9470b2b2a0db57361ac22df398e273af0e07a7a6dda8cf49686f"} Oct 08 00:29:46 crc kubenswrapper[4967]: I1008 00:29:46.072190 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-dlql4" event={"ID":"1f1c83bb-ff16-45ff-8fe5-e1b7e9d5166b","Type":"ContainerStarted","Data":"c26cd2eaa41a85ed9f7921f7bd04fb71bb49de85cc066ae5bd923881ad2abd7e"} Oct 08 00:29:46 crc kubenswrapper[4967]: I1008 00:29:46.079014 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/alertmanager-default-0" podStartSLOduration=15.271201163 podStartE2EDuration="29.07899678s" podCreationTimestamp="2025-10-08 00:29:17 +0000 UTC" firstStartedPulling="2025-10-08 00:29:31.930755872 +0000 UTC m=+1370.581290521" lastFinishedPulling="2025-10-08 00:29:45.738551479 +0000 UTC m=+1384.389086138" observedRunningTime="2025-10-08 00:29:46.076373013 +0000 UTC m=+1384.726907672" watchObservedRunningTime="2025-10-08 00:29:46.07899678 +0000 UTC m=+1384.729531429" Oct 08 00:29:46 crc kubenswrapper[4967]: W1008 00:29:46.175204 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod84dadc03_b18c_490c_adb2_9d3f605865d1.slice/crio-9cc208398b266e4ba70274efdd8674de7f2efbb1141e8305544ac95cb207b403 WatchSource:0}: Error finding container 9cc208398b266e4ba70274efdd8674de7f2efbb1141e8305544ac95cb207b403: Status 404 returned error can't find the container with id 9cc208398b266e4ba70274efdd8674de7f2efbb1141e8305544ac95cb207b403 Oct 08 00:29:46 crc kubenswrapper[4967]: I1008 00:29:46.177093 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-ceil-event-smartgateway-5d47c9955c-t24dp"] Oct 08 00:29:46 crc kubenswrapper[4967]: W1008 00:29:46.177546 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb3e181c2_79f9_47be_b4fd_f9f05599395f.slice/crio-c542a41ff7fdb0003da924f15b352a34eef7a466548714e819d286af377cd35e WatchSource:0}: Error finding container c542a41ff7fdb0003da924f15b352a34eef7a466548714e819d286af377cd35e: Status 404 returned error can't find the container with id c542a41ff7fdb0003da924f15b352a34eef7a466548714e819d286af377cd35e Oct 08 00:29:46 crc kubenswrapper[4967]: I1008 00:29:46.183210 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-coll-event-smartgateway-6f7464cc9d-bjhng"] Oct 08 00:29:47 crc kubenswrapper[4967]: I1008 00:29:47.079993 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-5d47c9955c-t24dp" event={"ID":"b3e181c2-79f9-47be-b4fd-f9f05599395f","Type":"ContainerStarted","Data":"b1ec56a770870b7a640e189bbd0bf76982a8e653f60d9b488b18045fb808c875"} Oct 08 00:29:47 crc kubenswrapper[4967]: I1008 00:29:47.080246 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-5d47c9955c-t24dp" event={"ID":"b3e181c2-79f9-47be-b4fd-f9f05599395f","Type":"ContainerStarted","Data":"c542a41ff7fdb0003da924f15b352a34eef7a466548714e819d286af377cd35e"} Oct 08 00:29:47 crc kubenswrapper[4967]: I1008 00:29:47.081270 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-event-smartgateway-6f7464cc9d-bjhng" event={"ID":"84dadc03-b18c-490c-adb2-9d3f605865d1","Type":"ContainerStarted","Data":"34accac679476fc42f9c5767708f217024880312fc26ff244aa579240f725b27"} Oct 08 00:29:47 crc kubenswrapper[4967]: I1008 00:29:47.081297 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-event-smartgateway-6f7464cc9d-bjhng" event={"ID":"84dadc03-b18c-490c-adb2-9d3f605865d1","Type":"ContainerStarted","Data":"9cc208398b266e4ba70274efdd8674de7f2efbb1141e8305544ac95cb207b403"} Oct 08 00:29:51 crc kubenswrapper[4967]: I1008 00:29:51.606539 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="service-telemetry/prometheus-default-0" Oct 08 00:29:51 crc kubenswrapper[4967]: I1008 00:29:51.658912 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="service-telemetry/prometheus-default-0" Oct 08 00:29:52 crc kubenswrapper[4967]: I1008 00:29:52.154946 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-rwnt8" event={"ID":"a9da825c-48d2-477d-b8c5-1ac878239c37","Type":"ContainerStarted","Data":"93d2bcc22aa555ce8a10311f4dcdc10891da7e34ea8631bb064e551dd5e97bb4"} Oct 08 00:29:52 crc kubenswrapper[4967]: I1008 00:29:52.155380 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-qcv8n" event={"ID":"ddba1803-397c-4daf-bc56-4857967229d1","Type":"ContainerStarted","Data":"870756f224961ccb97f3452a8a4da8ad8804807a0c59eca5047a00aacda2679c"} Oct 08 00:29:52 crc kubenswrapper[4967]: I1008 00:29:52.155402 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-dlql4" event={"ID":"1f1c83bb-ff16-45ff-8fe5-e1b7e9d5166b","Type":"ContainerStarted","Data":"d75b639e410e97879c9eca2b726c8b305b5c45f84a969d5a8963be497d173b75"} Oct 08 00:29:52 crc kubenswrapper[4967]: I1008 00:29:52.155422 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-5d47c9955c-t24dp" event={"ID":"b3e181c2-79f9-47be-b4fd-f9f05599395f","Type":"ContainerStarted","Data":"fa0554774066413d269384823a2fe24f1931fc2f590e704d5c0828acbe21a9b2"} Oct 08 00:29:52 crc kubenswrapper[4967]: I1008 00:29:52.157136 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-event-smartgateway-6f7464cc9d-bjhng" event={"ID":"84dadc03-b18c-490c-adb2-9d3f605865d1","Type":"ContainerStarted","Data":"111a4f75c824a9823f7d064d769a208b1e84defab38df2ae98835370c29b4042"} Oct 08 00:29:52 crc kubenswrapper[4967]: I1008 00:29:52.201362 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-rwnt8" podStartSLOduration=4.038797986 podStartE2EDuration="15.201334785s" podCreationTimestamp="2025-10-08 00:29:37 +0000 UTC" firstStartedPulling="2025-10-08 00:29:40.455889368 +0000 UTC m=+1379.106424017" lastFinishedPulling="2025-10-08 00:29:51.618426137 +0000 UTC m=+1390.268960816" observedRunningTime="2025-10-08 00:29:52.194794941 +0000 UTC m=+1390.845329620" watchObservedRunningTime="2025-10-08 00:29:52.201334785 +0000 UTC m=+1390.851869444" Oct 08 00:29:52 crc kubenswrapper[4967]: I1008 00:29:52.203819 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="service-telemetry/prometheus-default-0" Oct 08 00:29:52 crc kubenswrapper[4967]: I1008 00:29:52.218631 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-5d47c9955c-t24dp" podStartSLOduration=2.679072041 podStartE2EDuration="8.218586599s" podCreationTimestamp="2025-10-08 00:29:44 +0000 UTC" firstStartedPulling="2025-10-08 00:29:46.198441734 +0000 UTC m=+1384.848976373" lastFinishedPulling="2025-10-08 00:29:51.737956282 +0000 UTC m=+1390.388490931" observedRunningTime="2025-10-08 00:29:52.214728052 +0000 UTC m=+1390.865262721" watchObservedRunningTime="2025-10-08 00:29:52.218586599 +0000 UTC m=+1390.869121258" Oct 08 00:29:52 crc kubenswrapper[4967]: I1008 00:29:52.259849 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-dlql4" podStartSLOduration=7.496492463 podStartE2EDuration="20.259826876s" podCreationTimestamp="2025-10-08 00:29:32 +0000 UTC" firstStartedPulling="2025-10-08 00:29:38.860206563 +0000 UTC m=+1377.510741212" lastFinishedPulling="2025-10-08 00:29:51.623540946 +0000 UTC m=+1390.274075625" observedRunningTime="2025-10-08 00:29:52.250090061 +0000 UTC m=+1390.900624740" watchObservedRunningTime="2025-10-08 00:29:52.259826876 +0000 UTC m=+1390.910361525" Oct 08 00:29:52 crc kubenswrapper[4967]: I1008 00:29:52.273399 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-qcv8n" podStartSLOduration=9.274636555 podStartE2EDuration="22.273381937s" podCreationTimestamp="2025-10-08 00:29:30 +0000 UTC" firstStartedPulling="2025-10-08 00:29:38.687451059 +0000 UTC m=+1377.337985708" lastFinishedPulling="2025-10-08 00:29:51.686196441 +0000 UTC m=+1390.336731090" observedRunningTime="2025-10-08 00:29:52.268753281 +0000 UTC m=+1390.919287930" watchObservedRunningTime="2025-10-08 00:29:52.273381937 +0000 UTC m=+1390.923916576" Oct 08 00:29:52 crc kubenswrapper[4967]: I1008 00:29:52.292763 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-cloud1-coll-event-smartgateway-6f7464cc9d-bjhng" podStartSLOduration=2.717501466 podStartE2EDuration="8.292742744s" podCreationTimestamp="2025-10-08 00:29:44 +0000 UTC" firstStartedPulling="2025-10-08 00:29:46.177822945 +0000 UTC m=+1384.828357594" lastFinishedPulling="2025-10-08 00:29:51.753064223 +0000 UTC m=+1390.403598872" observedRunningTime="2025-10-08 00:29:52.289205875 +0000 UTC m=+1390.939740524" watchObservedRunningTime="2025-10-08 00:29:52.292742744 +0000 UTC m=+1390.943277393" Oct 08 00:29:55 crc kubenswrapper[4967]: I1008 00:29:55.875189 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/default-interconnect-68864d46cb-xxx8l"] Oct 08 00:29:55 crc kubenswrapper[4967]: I1008 00:29:55.876097 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="service-telemetry/default-interconnect-68864d46cb-xxx8l" podUID="e65fe86f-bcf1-41ac-b403-cc48da12ddb8" containerName="default-interconnect" containerID="cri-o://34e4a19674abb9a42557c3c696e1ec1784a807abe6c44aa3f5afdced1032a0c6" gracePeriod=30 Oct 08 00:29:56 crc kubenswrapper[4967]: I1008 00:29:56.190851 4967 generic.go:334] "Generic (PLEG): container finished" podID="ddba1803-397c-4daf-bc56-4857967229d1" containerID="3a11ae15580d9470b2b2a0db57361ac22df398e273af0e07a7a6dda8cf49686f" exitCode=0 Oct 08 00:29:56 crc kubenswrapper[4967]: I1008 00:29:56.190924 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-qcv8n" event={"ID":"ddba1803-397c-4daf-bc56-4857967229d1","Type":"ContainerDied","Data":"3a11ae15580d9470b2b2a0db57361ac22df398e273af0e07a7a6dda8cf49686f"} Oct 08 00:29:56 crc kubenswrapper[4967]: I1008 00:29:56.191567 4967 scope.go:117] "RemoveContainer" containerID="3a11ae15580d9470b2b2a0db57361ac22df398e273af0e07a7a6dda8cf49686f" Oct 08 00:29:56 crc kubenswrapper[4967]: I1008 00:29:56.201263 4967 generic.go:334] "Generic (PLEG): container finished" podID="1f1c83bb-ff16-45ff-8fe5-e1b7e9d5166b" containerID="c26cd2eaa41a85ed9f7921f7bd04fb71bb49de85cc066ae5bd923881ad2abd7e" exitCode=0 Oct 08 00:29:56 crc kubenswrapper[4967]: I1008 00:29:56.201403 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-dlql4" event={"ID":"1f1c83bb-ff16-45ff-8fe5-e1b7e9d5166b","Type":"ContainerDied","Data":"c26cd2eaa41a85ed9f7921f7bd04fb71bb49de85cc066ae5bd923881ad2abd7e"} Oct 08 00:29:56 crc kubenswrapper[4967]: I1008 00:29:56.202121 4967 scope.go:117] "RemoveContainer" containerID="c26cd2eaa41a85ed9f7921f7bd04fb71bb49de85cc066ae5bd923881ad2abd7e" Oct 08 00:29:56 crc kubenswrapper[4967]: I1008 00:29:56.208227 4967 generic.go:334] "Generic (PLEG): container finished" podID="e65fe86f-bcf1-41ac-b403-cc48da12ddb8" containerID="34e4a19674abb9a42557c3c696e1ec1784a807abe6c44aa3f5afdced1032a0c6" exitCode=0 Oct 08 00:29:56 crc kubenswrapper[4967]: I1008 00:29:56.208303 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-interconnect-68864d46cb-xxx8l" event={"ID":"e65fe86f-bcf1-41ac-b403-cc48da12ddb8","Type":"ContainerDied","Data":"34e4a19674abb9a42557c3c696e1ec1784a807abe6c44aa3f5afdced1032a0c6"} Oct 08 00:29:56 crc kubenswrapper[4967]: I1008 00:29:56.230332 4967 generic.go:334] "Generic (PLEG): container finished" podID="a9da825c-48d2-477d-b8c5-1ac878239c37" containerID="0526ba518406905f366079730b0681ac33f7b4d3625854c26e781db9e85f01d1" exitCode=0 Oct 08 00:29:56 crc kubenswrapper[4967]: I1008 00:29:56.230395 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-rwnt8" event={"ID":"a9da825c-48d2-477d-b8c5-1ac878239c37","Type":"ContainerDied","Data":"0526ba518406905f366079730b0681ac33f7b4d3625854c26e781db9e85f01d1"} Oct 08 00:29:56 crc kubenswrapper[4967]: I1008 00:29:56.231152 4967 scope.go:117] "RemoveContainer" containerID="0526ba518406905f366079730b0681ac33f7b4d3625854c26e781db9e85f01d1" Oct 08 00:29:56 crc kubenswrapper[4967]: I1008 00:29:56.356307 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-interconnect-68864d46cb-xxx8l" Oct 08 00:29:56 crc kubenswrapper[4967]: I1008 00:29:56.491017 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/e65fe86f-bcf1-41ac-b403-cc48da12ddb8-default-interconnect-openstack-credentials\") pod \"e65fe86f-bcf1-41ac-b403-cc48da12ddb8\" (UID: \"e65fe86f-bcf1-41ac-b403-cc48da12ddb8\") " Oct 08 00:29:56 crc kubenswrapper[4967]: I1008 00:29:56.491080 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-llgwb\" (UniqueName: \"kubernetes.io/projected/e65fe86f-bcf1-41ac-b403-cc48da12ddb8-kube-api-access-llgwb\") pod \"e65fe86f-bcf1-41ac-b403-cc48da12ddb8\" (UID: \"e65fe86f-bcf1-41ac-b403-cc48da12ddb8\") " Oct 08 00:29:56 crc kubenswrapper[4967]: I1008 00:29:56.491100 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/e65fe86f-bcf1-41ac-b403-cc48da12ddb8-default-interconnect-inter-router-credentials\") pod \"e65fe86f-bcf1-41ac-b403-cc48da12ddb8\" (UID: \"e65fe86f-bcf1-41ac-b403-cc48da12ddb8\") " Oct 08 00:29:56 crc kubenswrapper[4967]: I1008 00:29:56.491124 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/e65fe86f-bcf1-41ac-b403-cc48da12ddb8-sasl-users\") pod \"e65fe86f-bcf1-41ac-b403-cc48da12ddb8\" (UID: \"e65fe86f-bcf1-41ac-b403-cc48da12ddb8\") " Oct 08 00:29:56 crc kubenswrapper[4967]: I1008 00:29:56.491178 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/e65fe86f-bcf1-41ac-b403-cc48da12ddb8-sasl-config\") pod \"e65fe86f-bcf1-41ac-b403-cc48da12ddb8\" (UID: \"e65fe86f-bcf1-41ac-b403-cc48da12ddb8\") " Oct 08 00:29:56 crc kubenswrapper[4967]: I1008 00:29:56.491205 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/e65fe86f-bcf1-41ac-b403-cc48da12ddb8-default-interconnect-openstack-ca\") pod \"e65fe86f-bcf1-41ac-b403-cc48da12ddb8\" (UID: \"e65fe86f-bcf1-41ac-b403-cc48da12ddb8\") " Oct 08 00:29:56 crc kubenswrapper[4967]: I1008 00:29:56.491229 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/e65fe86f-bcf1-41ac-b403-cc48da12ddb8-default-interconnect-inter-router-ca\") pod \"e65fe86f-bcf1-41ac-b403-cc48da12ddb8\" (UID: \"e65fe86f-bcf1-41ac-b403-cc48da12ddb8\") " Oct 08 00:29:56 crc kubenswrapper[4967]: I1008 00:29:56.492842 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e65fe86f-bcf1-41ac-b403-cc48da12ddb8-sasl-config" (OuterVolumeSpecName: "sasl-config") pod "e65fe86f-bcf1-41ac-b403-cc48da12ddb8" (UID: "e65fe86f-bcf1-41ac-b403-cc48da12ddb8"). InnerVolumeSpecName "sasl-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:29:56 crc kubenswrapper[4967]: I1008 00:29:56.497023 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e65fe86f-bcf1-41ac-b403-cc48da12ddb8-default-interconnect-openstack-ca" (OuterVolumeSpecName: "default-interconnect-openstack-ca") pod "e65fe86f-bcf1-41ac-b403-cc48da12ddb8" (UID: "e65fe86f-bcf1-41ac-b403-cc48da12ddb8"). InnerVolumeSpecName "default-interconnect-openstack-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:29:56 crc kubenswrapper[4967]: I1008 00:29:56.497065 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e65fe86f-bcf1-41ac-b403-cc48da12ddb8-default-interconnect-openstack-credentials" (OuterVolumeSpecName: "default-interconnect-openstack-credentials") pod "e65fe86f-bcf1-41ac-b403-cc48da12ddb8" (UID: "e65fe86f-bcf1-41ac-b403-cc48da12ddb8"). InnerVolumeSpecName "default-interconnect-openstack-credentials". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:29:56 crc kubenswrapper[4967]: I1008 00:29:56.497252 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e65fe86f-bcf1-41ac-b403-cc48da12ddb8-kube-api-access-llgwb" (OuterVolumeSpecName: "kube-api-access-llgwb") pod "e65fe86f-bcf1-41ac-b403-cc48da12ddb8" (UID: "e65fe86f-bcf1-41ac-b403-cc48da12ddb8"). InnerVolumeSpecName "kube-api-access-llgwb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:29:56 crc kubenswrapper[4967]: I1008 00:29:56.497350 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e65fe86f-bcf1-41ac-b403-cc48da12ddb8-default-interconnect-inter-router-ca" (OuterVolumeSpecName: "default-interconnect-inter-router-ca") pod "e65fe86f-bcf1-41ac-b403-cc48da12ddb8" (UID: "e65fe86f-bcf1-41ac-b403-cc48da12ddb8"). InnerVolumeSpecName "default-interconnect-inter-router-ca". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:29:56 crc kubenswrapper[4967]: I1008 00:29:56.497398 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e65fe86f-bcf1-41ac-b403-cc48da12ddb8-default-interconnect-inter-router-credentials" (OuterVolumeSpecName: "default-interconnect-inter-router-credentials") pod "e65fe86f-bcf1-41ac-b403-cc48da12ddb8" (UID: "e65fe86f-bcf1-41ac-b403-cc48da12ddb8"). InnerVolumeSpecName "default-interconnect-inter-router-credentials". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:29:56 crc kubenswrapper[4967]: I1008 00:29:56.510903 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e65fe86f-bcf1-41ac-b403-cc48da12ddb8-sasl-users" (OuterVolumeSpecName: "sasl-users") pod "e65fe86f-bcf1-41ac-b403-cc48da12ddb8" (UID: "e65fe86f-bcf1-41ac-b403-cc48da12ddb8"). InnerVolumeSpecName "sasl-users". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:29:56 crc kubenswrapper[4967]: I1008 00:29:56.595619 4967 reconciler_common.go:293] "Volume detached for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/e65fe86f-bcf1-41ac-b403-cc48da12ddb8-default-interconnect-openstack-credentials\") on node \"crc\" DevicePath \"\"" Oct 08 00:29:56 crc kubenswrapper[4967]: I1008 00:29:56.595687 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-llgwb\" (UniqueName: \"kubernetes.io/projected/e65fe86f-bcf1-41ac-b403-cc48da12ddb8-kube-api-access-llgwb\") on node \"crc\" DevicePath \"\"" Oct 08 00:29:56 crc kubenswrapper[4967]: I1008 00:29:56.595703 4967 reconciler_common.go:293] "Volume detached for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/e65fe86f-bcf1-41ac-b403-cc48da12ddb8-default-interconnect-inter-router-credentials\") on node \"crc\" DevicePath \"\"" Oct 08 00:29:56 crc kubenswrapper[4967]: I1008 00:29:56.595715 4967 reconciler_common.go:293] "Volume detached for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/e65fe86f-bcf1-41ac-b403-cc48da12ddb8-sasl-users\") on node \"crc\" DevicePath \"\"" Oct 08 00:29:56 crc kubenswrapper[4967]: I1008 00:29:56.595727 4967 reconciler_common.go:293] "Volume detached for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/e65fe86f-bcf1-41ac-b403-cc48da12ddb8-sasl-config\") on node \"crc\" DevicePath \"\"" Oct 08 00:29:56 crc kubenswrapper[4967]: I1008 00:29:56.595738 4967 reconciler_common.go:293] "Volume detached for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/e65fe86f-bcf1-41ac-b403-cc48da12ddb8-default-interconnect-openstack-ca\") on node \"crc\" DevicePath \"\"" Oct 08 00:29:56 crc kubenswrapper[4967]: I1008 00:29:56.595753 4967 reconciler_common.go:293] "Volume detached for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/e65fe86f-bcf1-41ac-b403-cc48da12ddb8-default-interconnect-inter-router-ca\") on node \"crc\" DevicePath \"\"" Oct 08 00:29:56 crc kubenswrapper[4967]: I1008 00:29:56.907801 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/default-interconnect-68864d46cb-nb9sh"] Oct 08 00:29:56 crc kubenswrapper[4967]: E1008 00:29:56.908091 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e65fe86f-bcf1-41ac-b403-cc48da12ddb8" containerName="default-interconnect" Oct 08 00:29:56 crc kubenswrapper[4967]: I1008 00:29:56.908109 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="e65fe86f-bcf1-41ac-b403-cc48da12ddb8" containerName="default-interconnect" Oct 08 00:29:56 crc kubenswrapper[4967]: I1008 00:29:56.908246 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="e65fe86f-bcf1-41ac-b403-cc48da12ddb8" containerName="default-interconnect" Oct 08 00:29:56 crc kubenswrapper[4967]: I1008 00:29:56.908729 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-interconnect-68864d46cb-nb9sh" Oct 08 00:29:56 crc kubenswrapper[4967]: I1008 00:29:56.919114 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-interconnect-68864d46cb-nb9sh"] Oct 08 00:29:57 crc kubenswrapper[4967]: I1008 00:29:57.000772 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/94d8be0a-4986-4ff3-b327-b6da51cfdd64-sasl-config\") pod \"default-interconnect-68864d46cb-nb9sh\" (UID: \"94d8be0a-4986-4ff3-b327-b6da51cfdd64\") " pod="service-telemetry/default-interconnect-68864d46cb-nb9sh" Oct 08 00:29:57 crc kubenswrapper[4967]: I1008 00:29:57.000843 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/94d8be0a-4986-4ff3-b327-b6da51cfdd64-default-interconnect-inter-router-ca\") pod \"default-interconnect-68864d46cb-nb9sh\" (UID: \"94d8be0a-4986-4ff3-b327-b6da51cfdd64\") " pod="service-telemetry/default-interconnect-68864d46cb-nb9sh" Oct 08 00:29:57 crc kubenswrapper[4967]: I1008 00:29:57.000873 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/94d8be0a-4986-4ff3-b327-b6da51cfdd64-default-interconnect-openstack-credentials\") pod \"default-interconnect-68864d46cb-nb9sh\" (UID: \"94d8be0a-4986-4ff3-b327-b6da51cfdd64\") " pod="service-telemetry/default-interconnect-68864d46cb-nb9sh" Oct 08 00:29:57 crc kubenswrapper[4967]: I1008 00:29:57.000951 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/94d8be0a-4986-4ff3-b327-b6da51cfdd64-sasl-users\") pod \"default-interconnect-68864d46cb-nb9sh\" (UID: \"94d8be0a-4986-4ff3-b327-b6da51cfdd64\") " pod="service-telemetry/default-interconnect-68864d46cb-nb9sh" Oct 08 00:29:57 crc kubenswrapper[4967]: I1008 00:29:57.001017 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/94d8be0a-4986-4ff3-b327-b6da51cfdd64-default-interconnect-openstack-ca\") pod \"default-interconnect-68864d46cb-nb9sh\" (UID: \"94d8be0a-4986-4ff3-b327-b6da51cfdd64\") " pod="service-telemetry/default-interconnect-68864d46cb-nb9sh" Oct 08 00:29:57 crc kubenswrapper[4967]: I1008 00:29:57.001072 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-46dxv\" (UniqueName: \"kubernetes.io/projected/94d8be0a-4986-4ff3-b327-b6da51cfdd64-kube-api-access-46dxv\") pod \"default-interconnect-68864d46cb-nb9sh\" (UID: \"94d8be0a-4986-4ff3-b327-b6da51cfdd64\") " pod="service-telemetry/default-interconnect-68864d46cb-nb9sh" Oct 08 00:29:57 crc kubenswrapper[4967]: I1008 00:29:57.001112 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/94d8be0a-4986-4ff3-b327-b6da51cfdd64-default-interconnect-inter-router-credentials\") pod \"default-interconnect-68864d46cb-nb9sh\" (UID: \"94d8be0a-4986-4ff3-b327-b6da51cfdd64\") " pod="service-telemetry/default-interconnect-68864d46cb-nb9sh" Oct 08 00:29:57 crc kubenswrapper[4967]: I1008 00:29:57.102704 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/94d8be0a-4986-4ff3-b327-b6da51cfdd64-sasl-users\") pod \"default-interconnect-68864d46cb-nb9sh\" (UID: \"94d8be0a-4986-4ff3-b327-b6da51cfdd64\") " pod="service-telemetry/default-interconnect-68864d46cb-nb9sh" Oct 08 00:29:57 crc kubenswrapper[4967]: I1008 00:29:57.102761 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/94d8be0a-4986-4ff3-b327-b6da51cfdd64-default-interconnect-openstack-ca\") pod \"default-interconnect-68864d46cb-nb9sh\" (UID: \"94d8be0a-4986-4ff3-b327-b6da51cfdd64\") " pod="service-telemetry/default-interconnect-68864d46cb-nb9sh" Oct 08 00:29:57 crc kubenswrapper[4967]: I1008 00:29:57.102796 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-46dxv\" (UniqueName: \"kubernetes.io/projected/94d8be0a-4986-4ff3-b327-b6da51cfdd64-kube-api-access-46dxv\") pod \"default-interconnect-68864d46cb-nb9sh\" (UID: \"94d8be0a-4986-4ff3-b327-b6da51cfdd64\") " pod="service-telemetry/default-interconnect-68864d46cb-nb9sh" Oct 08 00:29:57 crc kubenswrapper[4967]: I1008 00:29:57.102820 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/94d8be0a-4986-4ff3-b327-b6da51cfdd64-default-interconnect-inter-router-credentials\") pod \"default-interconnect-68864d46cb-nb9sh\" (UID: \"94d8be0a-4986-4ff3-b327-b6da51cfdd64\") " pod="service-telemetry/default-interconnect-68864d46cb-nb9sh" Oct 08 00:29:57 crc kubenswrapper[4967]: I1008 00:29:57.102853 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/94d8be0a-4986-4ff3-b327-b6da51cfdd64-sasl-config\") pod \"default-interconnect-68864d46cb-nb9sh\" (UID: \"94d8be0a-4986-4ff3-b327-b6da51cfdd64\") " pod="service-telemetry/default-interconnect-68864d46cb-nb9sh" Oct 08 00:29:57 crc kubenswrapper[4967]: I1008 00:29:57.102880 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/94d8be0a-4986-4ff3-b327-b6da51cfdd64-default-interconnect-inter-router-ca\") pod \"default-interconnect-68864d46cb-nb9sh\" (UID: \"94d8be0a-4986-4ff3-b327-b6da51cfdd64\") " pod="service-telemetry/default-interconnect-68864d46cb-nb9sh" Oct 08 00:29:57 crc kubenswrapper[4967]: I1008 00:29:57.102899 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/94d8be0a-4986-4ff3-b327-b6da51cfdd64-default-interconnect-openstack-credentials\") pod \"default-interconnect-68864d46cb-nb9sh\" (UID: \"94d8be0a-4986-4ff3-b327-b6da51cfdd64\") " pod="service-telemetry/default-interconnect-68864d46cb-nb9sh" Oct 08 00:29:57 crc kubenswrapper[4967]: I1008 00:29:57.103989 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/94d8be0a-4986-4ff3-b327-b6da51cfdd64-sasl-config\") pod \"default-interconnect-68864d46cb-nb9sh\" (UID: \"94d8be0a-4986-4ff3-b327-b6da51cfdd64\") " pod="service-telemetry/default-interconnect-68864d46cb-nb9sh" Oct 08 00:29:57 crc kubenswrapper[4967]: I1008 00:29:57.107177 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/94d8be0a-4986-4ff3-b327-b6da51cfdd64-default-interconnect-openstack-credentials\") pod \"default-interconnect-68864d46cb-nb9sh\" (UID: \"94d8be0a-4986-4ff3-b327-b6da51cfdd64\") " pod="service-telemetry/default-interconnect-68864d46cb-nb9sh" Oct 08 00:29:57 crc kubenswrapper[4967]: I1008 00:29:57.107257 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/94d8be0a-4986-4ff3-b327-b6da51cfdd64-default-interconnect-openstack-ca\") pod \"default-interconnect-68864d46cb-nb9sh\" (UID: \"94d8be0a-4986-4ff3-b327-b6da51cfdd64\") " pod="service-telemetry/default-interconnect-68864d46cb-nb9sh" Oct 08 00:29:57 crc kubenswrapper[4967]: I1008 00:29:57.115505 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/94d8be0a-4986-4ff3-b327-b6da51cfdd64-sasl-users\") pod \"default-interconnect-68864d46cb-nb9sh\" (UID: \"94d8be0a-4986-4ff3-b327-b6da51cfdd64\") " pod="service-telemetry/default-interconnect-68864d46cb-nb9sh" Oct 08 00:29:57 crc kubenswrapper[4967]: I1008 00:29:57.115697 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/94d8be0a-4986-4ff3-b327-b6da51cfdd64-default-interconnect-inter-router-credentials\") pod \"default-interconnect-68864d46cb-nb9sh\" (UID: \"94d8be0a-4986-4ff3-b327-b6da51cfdd64\") " pod="service-telemetry/default-interconnect-68864d46cb-nb9sh" Oct 08 00:29:57 crc kubenswrapper[4967]: I1008 00:29:57.116192 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/94d8be0a-4986-4ff3-b327-b6da51cfdd64-default-interconnect-inter-router-ca\") pod \"default-interconnect-68864d46cb-nb9sh\" (UID: \"94d8be0a-4986-4ff3-b327-b6da51cfdd64\") " pod="service-telemetry/default-interconnect-68864d46cb-nb9sh" Oct 08 00:29:57 crc kubenswrapper[4967]: I1008 00:29:57.123654 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-46dxv\" (UniqueName: \"kubernetes.io/projected/94d8be0a-4986-4ff3-b327-b6da51cfdd64-kube-api-access-46dxv\") pod \"default-interconnect-68864d46cb-nb9sh\" (UID: \"94d8be0a-4986-4ff3-b327-b6da51cfdd64\") " pod="service-telemetry/default-interconnect-68864d46cb-nb9sh" Oct 08 00:29:57 crc kubenswrapper[4967]: I1008 00:29:57.225090 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-interconnect-68864d46cb-nb9sh" Oct 08 00:29:57 crc kubenswrapper[4967]: I1008 00:29:57.252956 4967 generic.go:334] "Generic (PLEG): container finished" podID="84dadc03-b18c-490c-adb2-9d3f605865d1" containerID="34accac679476fc42f9c5767708f217024880312fc26ff244aa579240f725b27" exitCode=0 Oct 08 00:29:57 crc kubenswrapper[4967]: I1008 00:29:57.253048 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-event-smartgateway-6f7464cc9d-bjhng" event={"ID":"84dadc03-b18c-490c-adb2-9d3f605865d1","Type":"ContainerDied","Data":"34accac679476fc42f9c5767708f217024880312fc26ff244aa579240f725b27"} Oct 08 00:29:57 crc kubenswrapper[4967]: I1008 00:29:57.254024 4967 scope.go:117] "RemoveContainer" containerID="34accac679476fc42f9c5767708f217024880312fc26ff244aa579240f725b27" Oct 08 00:29:57 crc kubenswrapper[4967]: I1008 00:29:57.262941 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-interconnect-68864d46cb-xxx8l" event={"ID":"e65fe86f-bcf1-41ac-b403-cc48da12ddb8","Type":"ContainerDied","Data":"061646feae812d3f37b1837e7d72df04d4d30adffc838c77a08d0277a1c8f863"} Oct 08 00:29:57 crc kubenswrapper[4967]: I1008 00:29:57.262993 4967 scope.go:117] "RemoveContainer" containerID="34e4a19674abb9a42557c3c696e1ec1784a807abe6c44aa3f5afdced1032a0c6" Oct 08 00:29:57 crc kubenswrapper[4967]: I1008 00:29:57.263206 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-interconnect-68864d46cb-xxx8l" Oct 08 00:29:57 crc kubenswrapper[4967]: I1008 00:29:57.292242 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-rwnt8" event={"ID":"a9da825c-48d2-477d-b8c5-1ac878239c37","Type":"ContainerStarted","Data":"45616f6aeab51b4f086393063c6d90a2738cdb3308a9acd32ac19dfa9003472d"} Oct 08 00:29:57 crc kubenswrapper[4967]: I1008 00:29:57.304953 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-qcv8n" event={"ID":"ddba1803-397c-4daf-bc56-4857967229d1","Type":"ContainerStarted","Data":"172a90c9893e70774c41d3027c26eac57aca1df3d0c370d72d1a1e7bd879cb82"} Oct 08 00:29:57 crc kubenswrapper[4967]: I1008 00:29:57.314465 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-dlql4" event={"ID":"1f1c83bb-ff16-45ff-8fe5-e1b7e9d5166b","Type":"ContainerStarted","Data":"5708a0875e9b4587f5e7e31ae66f8fa5ab5140fa37e4d1c8f1c2a9aceef79143"} Oct 08 00:29:57 crc kubenswrapper[4967]: I1008 00:29:57.375659 4967 generic.go:334] "Generic (PLEG): container finished" podID="b3e181c2-79f9-47be-b4fd-f9f05599395f" containerID="b1ec56a770870b7a640e189bbd0bf76982a8e653f60d9b488b18045fb808c875" exitCode=0 Oct 08 00:29:57 crc kubenswrapper[4967]: I1008 00:29:57.375718 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-5d47c9955c-t24dp" event={"ID":"b3e181c2-79f9-47be-b4fd-f9f05599395f","Type":"ContainerDied","Data":"b1ec56a770870b7a640e189bbd0bf76982a8e653f60d9b488b18045fb808c875"} Oct 08 00:29:57 crc kubenswrapper[4967]: I1008 00:29:57.376630 4967 scope.go:117] "RemoveContainer" containerID="b1ec56a770870b7a640e189bbd0bf76982a8e653f60d9b488b18045fb808c875" Oct 08 00:29:57 crc kubenswrapper[4967]: I1008 00:29:57.390551 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/default-interconnect-68864d46cb-xxx8l"] Oct 08 00:29:57 crc kubenswrapper[4967]: I1008 00:29:57.403911 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["service-telemetry/default-interconnect-68864d46cb-xxx8l"] Oct 08 00:29:57 crc kubenswrapper[4967]: I1008 00:29:57.619140 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-interconnect-68864d46cb-nb9sh"] Oct 08 00:29:58 crc kubenswrapper[4967]: I1008 00:29:58.125837 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e65fe86f-bcf1-41ac-b403-cc48da12ddb8" path="/var/lib/kubelet/pods/e65fe86f-bcf1-41ac-b403-cc48da12ddb8/volumes" Oct 08 00:29:58 crc kubenswrapper[4967]: I1008 00:29:58.394392 4967 generic.go:334] "Generic (PLEG): container finished" podID="a9da825c-48d2-477d-b8c5-1ac878239c37" containerID="45616f6aeab51b4f086393063c6d90a2738cdb3308a9acd32ac19dfa9003472d" exitCode=0 Oct 08 00:29:58 crc kubenswrapper[4967]: I1008 00:29:58.394453 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-rwnt8" event={"ID":"a9da825c-48d2-477d-b8c5-1ac878239c37","Type":"ContainerDied","Data":"45616f6aeab51b4f086393063c6d90a2738cdb3308a9acd32ac19dfa9003472d"} Oct 08 00:29:58 crc kubenswrapper[4967]: I1008 00:29:58.394493 4967 scope.go:117] "RemoveContainer" containerID="0526ba518406905f366079730b0681ac33f7b4d3625854c26e781db9e85f01d1" Oct 08 00:29:58 crc kubenswrapper[4967]: I1008 00:29:58.395128 4967 scope.go:117] "RemoveContainer" containerID="45616f6aeab51b4f086393063c6d90a2738cdb3308a9acd32ac19dfa9003472d" Oct 08 00:29:58 crc kubenswrapper[4967]: E1008 00:29:58.395346 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bridge\" with CrashLoopBackOff: \"back-off 10s restarting failed container=bridge pod=default-cloud1-sens-meter-smartgateway-5759b4d97-rwnt8_service-telemetry(a9da825c-48d2-477d-b8c5-1ac878239c37)\"" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-rwnt8" podUID="a9da825c-48d2-477d-b8c5-1ac878239c37" Oct 08 00:29:58 crc kubenswrapper[4967]: I1008 00:29:58.397447 4967 generic.go:334] "Generic (PLEG): container finished" podID="ddba1803-397c-4daf-bc56-4857967229d1" containerID="172a90c9893e70774c41d3027c26eac57aca1df3d0c370d72d1a1e7bd879cb82" exitCode=0 Oct 08 00:29:58 crc kubenswrapper[4967]: I1008 00:29:58.397464 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-qcv8n" event={"ID":"ddba1803-397c-4daf-bc56-4857967229d1","Type":"ContainerDied","Data":"172a90c9893e70774c41d3027c26eac57aca1df3d0c370d72d1a1e7bd879cb82"} Oct 08 00:29:58 crc kubenswrapper[4967]: I1008 00:29:58.397968 4967 scope.go:117] "RemoveContainer" containerID="172a90c9893e70774c41d3027c26eac57aca1df3d0c370d72d1a1e7bd879cb82" Oct 08 00:29:58 crc kubenswrapper[4967]: E1008 00:29:58.398141 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bridge\" with CrashLoopBackOff: \"back-off 10s restarting failed container=bridge pod=default-cloud1-coll-meter-smartgateway-7cd87f9766-qcv8n_service-telemetry(ddba1803-397c-4daf-bc56-4857967229d1)\"" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-qcv8n" podUID="ddba1803-397c-4daf-bc56-4857967229d1" Oct 08 00:29:58 crc kubenswrapper[4967]: I1008 00:29:58.398931 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-interconnect-68864d46cb-nb9sh" event={"ID":"94d8be0a-4986-4ff3-b327-b6da51cfdd64","Type":"ContainerStarted","Data":"d62397be4ae6e453c3e85d1e0a0b3a8e05f242d0aa1375aa8a457fff066cbe3e"} Oct 08 00:29:58 crc kubenswrapper[4967]: I1008 00:29:58.398990 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-interconnect-68864d46cb-nb9sh" event={"ID":"94d8be0a-4986-4ff3-b327-b6da51cfdd64","Type":"ContainerStarted","Data":"cb34cac0074a94901749f93149c828490e95c4fcc77d856fdf62fa0b1d181cfc"} Oct 08 00:29:58 crc kubenswrapper[4967]: I1008 00:29:58.401855 4967 generic.go:334] "Generic (PLEG): container finished" podID="1f1c83bb-ff16-45ff-8fe5-e1b7e9d5166b" containerID="5708a0875e9b4587f5e7e31ae66f8fa5ab5140fa37e4d1c8f1c2a9aceef79143" exitCode=0 Oct 08 00:29:58 crc kubenswrapper[4967]: I1008 00:29:58.402092 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-dlql4" event={"ID":"1f1c83bb-ff16-45ff-8fe5-e1b7e9d5166b","Type":"ContainerDied","Data":"5708a0875e9b4587f5e7e31ae66f8fa5ab5140fa37e4d1c8f1c2a9aceef79143"} Oct 08 00:29:58 crc kubenswrapper[4967]: I1008 00:29:58.403396 4967 scope.go:117] "RemoveContainer" containerID="5708a0875e9b4587f5e7e31ae66f8fa5ab5140fa37e4d1c8f1c2a9aceef79143" Oct 08 00:29:58 crc kubenswrapper[4967]: E1008 00:29:58.403994 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bridge\" with CrashLoopBackOff: \"back-off 10s restarting failed container=bridge pod=default-cloud1-ceil-meter-smartgateway-57948895dc-dlql4_service-telemetry(1f1c83bb-ff16-45ff-8fe5-e1b7e9d5166b)\"" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-dlql4" podUID="1f1c83bb-ff16-45ff-8fe5-e1b7e9d5166b" Oct 08 00:29:58 crc kubenswrapper[4967]: I1008 00:29:58.404883 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-5d47c9955c-t24dp" event={"ID":"b3e181c2-79f9-47be-b4fd-f9f05599395f","Type":"ContainerStarted","Data":"b7d19b7c1b6d22b5b39b4ecccaa58ab4f49169900f0d047f09cab046927a8037"} Oct 08 00:29:58 crc kubenswrapper[4967]: I1008 00:29:58.406774 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-event-smartgateway-6f7464cc9d-bjhng" event={"ID":"84dadc03-b18c-490c-adb2-9d3f605865d1","Type":"ContainerStarted","Data":"71734357db2a83fc820ed005863935126520931ebbe4df8c40ea34276b7831f9"} Oct 08 00:29:58 crc kubenswrapper[4967]: I1008 00:29:58.457411 4967 scope.go:117] "RemoveContainer" containerID="3a11ae15580d9470b2b2a0db57361ac22df398e273af0e07a7a6dda8cf49686f" Oct 08 00:29:58 crc kubenswrapper[4967]: I1008 00:29:58.516237 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-interconnect-68864d46cb-nb9sh" podStartSLOduration=3.516217502 podStartE2EDuration="3.516217502s" podCreationTimestamp="2025-10-08 00:29:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:29:58.477997361 +0000 UTC m=+1397.128532010" watchObservedRunningTime="2025-10-08 00:29:58.516217502 +0000 UTC m=+1397.166752151" Oct 08 00:29:58 crc kubenswrapper[4967]: I1008 00:29:58.527948 4967 scope.go:117] "RemoveContainer" containerID="c26cd2eaa41a85ed9f7921f7bd04fb71bb49de85cc066ae5bd923881ad2abd7e" Oct 08 00:29:58 crc kubenswrapper[4967]: I1008 00:29:58.739043 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-25rkp"] Oct 08 00:29:58 crc kubenswrapper[4967]: I1008 00:29:58.740570 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-25rkp" Oct 08 00:29:58 crc kubenswrapper[4967]: I1008 00:29:58.746214 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-25rkp"] Oct 08 00:29:58 crc kubenswrapper[4967]: I1008 00:29:58.832582 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f8e39c1-cde6-4742-9d2a-cfbb47b5a14b-catalog-content\") pod \"redhat-operators-25rkp\" (UID: \"5f8e39c1-cde6-4742-9d2a-cfbb47b5a14b\") " pod="openshift-marketplace/redhat-operators-25rkp" Oct 08 00:29:58 crc kubenswrapper[4967]: I1008 00:29:58.832655 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cbjl5\" (UniqueName: \"kubernetes.io/projected/5f8e39c1-cde6-4742-9d2a-cfbb47b5a14b-kube-api-access-cbjl5\") pod \"redhat-operators-25rkp\" (UID: \"5f8e39c1-cde6-4742-9d2a-cfbb47b5a14b\") " pod="openshift-marketplace/redhat-operators-25rkp" Oct 08 00:29:58 crc kubenswrapper[4967]: I1008 00:29:58.832724 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f8e39c1-cde6-4742-9d2a-cfbb47b5a14b-utilities\") pod \"redhat-operators-25rkp\" (UID: \"5f8e39c1-cde6-4742-9d2a-cfbb47b5a14b\") " pod="openshift-marketplace/redhat-operators-25rkp" Oct 08 00:29:58 crc kubenswrapper[4967]: I1008 00:29:58.934054 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f8e39c1-cde6-4742-9d2a-cfbb47b5a14b-catalog-content\") pod \"redhat-operators-25rkp\" (UID: \"5f8e39c1-cde6-4742-9d2a-cfbb47b5a14b\") " pod="openshift-marketplace/redhat-operators-25rkp" Oct 08 00:29:58 crc kubenswrapper[4967]: I1008 00:29:58.934127 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cbjl5\" (UniqueName: \"kubernetes.io/projected/5f8e39c1-cde6-4742-9d2a-cfbb47b5a14b-kube-api-access-cbjl5\") pod \"redhat-operators-25rkp\" (UID: \"5f8e39c1-cde6-4742-9d2a-cfbb47b5a14b\") " pod="openshift-marketplace/redhat-operators-25rkp" Oct 08 00:29:58 crc kubenswrapper[4967]: I1008 00:29:58.934197 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f8e39c1-cde6-4742-9d2a-cfbb47b5a14b-utilities\") pod \"redhat-operators-25rkp\" (UID: \"5f8e39c1-cde6-4742-9d2a-cfbb47b5a14b\") " pod="openshift-marketplace/redhat-operators-25rkp" Oct 08 00:29:58 crc kubenswrapper[4967]: I1008 00:29:58.934773 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f8e39c1-cde6-4742-9d2a-cfbb47b5a14b-utilities\") pod \"redhat-operators-25rkp\" (UID: \"5f8e39c1-cde6-4742-9d2a-cfbb47b5a14b\") " pod="openshift-marketplace/redhat-operators-25rkp" Oct 08 00:29:58 crc kubenswrapper[4967]: I1008 00:29:58.935124 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f8e39c1-cde6-4742-9d2a-cfbb47b5a14b-catalog-content\") pod \"redhat-operators-25rkp\" (UID: \"5f8e39c1-cde6-4742-9d2a-cfbb47b5a14b\") " pod="openshift-marketplace/redhat-operators-25rkp" Oct 08 00:29:58 crc kubenswrapper[4967]: I1008 00:29:58.954918 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cbjl5\" (UniqueName: \"kubernetes.io/projected/5f8e39c1-cde6-4742-9d2a-cfbb47b5a14b-kube-api-access-cbjl5\") pod \"redhat-operators-25rkp\" (UID: \"5f8e39c1-cde6-4742-9d2a-cfbb47b5a14b\") " pod="openshift-marketplace/redhat-operators-25rkp" Oct 08 00:29:59 crc kubenswrapper[4967]: I1008 00:29:59.066462 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-25rkp" Oct 08 00:29:59 crc kubenswrapper[4967]: I1008 00:29:59.425836 4967 generic.go:334] "Generic (PLEG): container finished" podID="b3e181c2-79f9-47be-b4fd-f9f05599395f" containerID="b7d19b7c1b6d22b5b39b4ecccaa58ab4f49169900f0d047f09cab046927a8037" exitCode=0 Oct 08 00:29:59 crc kubenswrapper[4967]: I1008 00:29:59.426195 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-5d47c9955c-t24dp" event={"ID":"b3e181c2-79f9-47be-b4fd-f9f05599395f","Type":"ContainerDied","Data":"b7d19b7c1b6d22b5b39b4ecccaa58ab4f49169900f0d047f09cab046927a8037"} Oct 08 00:29:59 crc kubenswrapper[4967]: I1008 00:29:59.426229 4967 scope.go:117] "RemoveContainer" containerID="b1ec56a770870b7a640e189bbd0bf76982a8e653f60d9b488b18045fb808c875" Oct 08 00:29:59 crc kubenswrapper[4967]: I1008 00:29:59.426694 4967 scope.go:117] "RemoveContainer" containerID="b7d19b7c1b6d22b5b39b4ecccaa58ab4f49169900f0d047f09cab046927a8037" Oct 08 00:29:59 crc kubenswrapper[4967]: E1008 00:29:59.426946 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bridge\" with CrashLoopBackOff: \"back-off 10s restarting failed container=bridge pod=default-cloud1-ceil-event-smartgateway-5d47c9955c-t24dp_service-telemetry(b3e181c2-79f9-47be-b4fd-f9f05599395f)\"" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-5d47c9955c-t24dp" podUID="b3e181c2-79f9-47be-b4fd-f9f05599395f" Oct 08 00:29:59 crc kubenswrapper[4967]: I1008 00:29:59.435794 4967 generic.go:334] "Generic (PLEG): container finished" podID="84dadc03-b18c-490c-adb2-9d3f605865d1" containerID="71734357db2a83fc820ed005863935126520931ebbe4df8c40ea34276b7831f9" exitCode=0 Oct 08 00:29:59 crc kubenswrapper[4967]: I1008 00:29:59.435885 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-event-smartgateway-6f7464cc9d-bjhng" event={"ID":"84dadc03-b18c-490c-adb2-9d3f605865d1","Type":"ContainerDied","Data":"71734357db2a83fc820ed005863935126520931ebbe4df8c40ea34276b7831f9"} Oct 08 00:29:59 crc kubenswrapper[4967]: I1008 00:29:59.436525 4967 scope.go:117] "RemoveContainer" containerID="71734357db2a83fc820ed005863935126520931ebbe4df8c40ea34276b7831f9" Oct 08 00:29:59 crc kubenswrapper[4967]: E1008 00:29:59.436807 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bridge\" with CrashLoopBackOff: \"back-off 10s restarting failed container=bridge pod=default-cloud1-coll-event-smartgateway-6f7464cc9d-bjhng_service-telemetry(84dadc03-b18c-490c-adb2-9d3f605865d1)\"" pod="service-telemetry/default-cloud1-coll-event-smartgateway-6f7464cc9d-bjhng" podUID="84dadc03-b18c-490c-adb2-9d3f605865d1" Oct 08 00:29:59 crc kubenswrapper[4967]: I1008 00:29:59.470159 4967 scope.go:117] "RemoveContainer" containerID="34accac679476fc42f9c5767708f217024880312fc26ff244aa579240f725b27" Oct 08 00:29:59 crc kubenswrapper[4967]: I1008 00:29:59.502626 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-25rkp"] Oct 08 00:29:59 crc kubenswrapper[4967]: W1008 00:29:59.512711 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5f8e39c1_cde6_4742_9d2a_cfbb47b5a14b.slice/crio-e0a8dca80f5e3eb594da12a03ca70d682ec48a86422c9c9bf19a6126a8dc81c6 WatchSource:0}: Error finding container e0a8dca80f5e3eb594da12a03ca70d682ec48a86422c9c9bf19a6126a8dc81c6: Status 404 returned error can't find the container with id e0a8dca80f5e3eb594da12a03ca70d682ec48a86422c9c9bf19a6126a8dc81c6 Oct 08 00:30:00 crc kubenswrapper[4967]: I1008 00:30:00.144843 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331390-cv6kh"] Oct 08 00:30:00 crc kubenswrapper[4967]: I1008 00:30:00.145952 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331390-cv6kh" Oct 08 00:30:00 crc kubenswrapper[4967]: I1008 00:30:00.152221 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 08 00:30:00 crc kubenswrapper[4967]: I1008 00:30:00.152880 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 08 00:30:00 crc kubenswrapper[4967]: I1008 00:30:00.158436 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331390-cv6kh"] Oct 08 00:30:00 crc kubenswrapper[4967]: I1008 00:30:00.254366 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kskkw\" (UniqueName: \"kubernetes.io/projected/8c219dd4-8db4-4015-8df0-ab4989c5be71-kube-api-access-kskkw\") pod \"collect-profiles-29331390-cv6kh\" (UID: \"8c219dd4-8db4-4015-8df0-ab4989c5be71\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331390-cv6kh" Oct 08 00:30:00 crc kubenswrapper[4967]: I1008 00:30:00.254488 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8c219dd4-8db4-4015-8df0-ab4989c5be71-config-volume\") pod \"collect-profiles-29331390-cv6kh\" (UID: \"8c219dd4-8db4-4015-8df0-ab4989c5be71\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331390-cv6kh" Oct 08 00:30:00 crc kubenswrapper[4967]: I1008 00:30:00.254523 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8c219dd4-8db4-4015-8df0-ab4989c5be71-secret-volume\") pod \"collect-profiles-29331390-cv6kh\" (UID: \"8c219dd4-8db4-4015-8df0-ab4989c5be71\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331390-cv6kh" Oct 08 00:30:00 crc kubenswrapper[4967]: I1008 00:30:00.356385 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kskkw\" (UniqueName: \"kubernetes.io/projected/8c219dd4-8db4-4015-8df0-ab4989c5be71-kube-api-access-kskkw\") pod \"collect-profiles-29331390-cv6kh\" (UID: \"8c219dd4-8db4-4015-8df0-ab4989c5be71\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331390-cv6kh" Oct 08 00:30:00 crc kubenswrapper[4967]: I1008 00:30:00.356477 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8c219dd4-8db4-4015-8df0-ab4989c5be71-config-volume\") pod \"collect-profiles-29331390-cv6kh\" (UID: \"8c219dd4-8db4-4015-8df0-ab4989c5be71\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331390-cv6kh" Oct 08 00:30:00 crc kubenswrapper[4967]: I1008 00:30:00.356503 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8c219dd4-8db4-4015-8df0-ab4989c5be71-secret-volume\") pod \"collect-profiles-29331390-cv6kh\" (UID: \"8c219dd4-8db4-4015-8df0-ab4989c5be71\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331390-cv6kh" Oct 08 00:30:00 crc kubenswrapper[4967]: I1008 00:30:00.358043 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8c219dd4-8db4-4015-8df0-ab4989c5be71-config-volume\") pod \"collect-profiles-29331390-cv6kh\" (UID: \"8c219dd4-8db4-4015-8df0-ab4989c5be71\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331390-cv6kh" Oct 08 00:30:00 crc kubenswrapper[4967]: I1008 00:30:00.374803 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kskkw\" (UniqueName: \"kubernetes.io/projected/8c219dd4-8db4-4015-8df0-ab4989c5be71-kube-api-access-kskkw\") pod \"collect-profiles-29331390-cv6kh\" (UID: \"8c219dd4-8db4-4015-8df0-ab4989c5be71\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331390-cv6kh" Oct 08 00:30:00 crc kubenswrapper[4967]: I1008 00:30:00.379988 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8c219dd4-8db4-4015-8df0-ab4989c5be71-secret-volume\") pod \"collect-profiles-29331390-cv6kh\" (UID: \"8c219dd4-8db4-4015-8df0-ab4989c5be71\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29331390-cv6kh" Oct 08 00:30:00 crc kubenswrapper[4967]: I1008 00:30:00.463032 4967 generic.go:334] "Generic (PLEG): container finished" podID="5f8e39c1-cde6-4742-9d2a-cfbb47b5a14b" containerID="ba88b4442e2df98ac6775b84a3453e241f2e9443a9cc6b61d2442ab80a2cc30c" exitCode=0 Oct 08 00:30:00 crc kubenswrapper[4967]: I1008 00:30:00.463160 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-25rkp" event={"ID":"5f8e39c1-cde6-4742-9d2a-cfbb47b5a14b","Type":"ContainerDied","Data":"ba88b4442e2df98ac6775b84a3453e241f2e9443a9cc6b61d2442ab80a2cc30c"} Oct 08 00:30:00 crc kubenswrapper[4967]: I1008 00:30:00.463207 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-25rkp" event={"ID":"5f8e39c1-cde6-4742-9d2a-cfbb47b5a14b","Type":"ContainerStarted","Data":"e0a8dca80f5e3eb594da12a03ca70d682ec48a86422c9c9bf19a6126a8dc81c6"} Oct 08 00:30:00 crc kubenswrapper[4967]: I1008 00:30:00.469451 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331390-cv6kh" Oct 08 00:30:00 crc kubenswrapper[4967]: I1008 00:30:00.902882 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29331390-cv6kh"] Oct 08 00:30:01 crc kubenswrapper[4967]: I1008 00:30:01.476344 4967 generic.go:334] "Generic (PLEG): container finished" podID="8c219dd4-8db4-4015-8df0-ab4989c5be71" containerID="4ddcf4ec25ca1b25591da6d24f1e4127e2f3beae7bc06a9140f095a2957f84a7" exitCode=0 Oct 08 00:30:01 crc kubenswrapper[4967]: I1008 00:30:01.476377 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331390-cv6kh" event={"ID":"8c219dd4-8db4-4015-8df0-ab4989c5be71","Type":"ContainerDied","Data":"4ddcf4ec25ca1b25591da6d24f1e4127e2f3beae7bc06a9140f095a2957f84a7"} Oct 08 00:30:01 crc kubenswrapper[4967]: I1008 00:30:01.476575 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331390-cv6kh" event={"ID":"8c219dd4-8db4-4015-8df0-ab4989c5be71","Type":"ContainerStarted","Data":"ca71e630d09e321aff646c4edbbd124bcdfc8aab41d137c358f38348b6434124"} Oct 08 00:30:02 crc kubenswrapper[4967]: I1008 00:30:02.485588 4967 generic.go:334] "Generic (PLEG): container finished" podID="5f8e39c1-cde6-4742-9d2a-cfbb47b5a14b" containerID="56abe64eac56578021cf8baf3775ab7018104e2a15d62306943d047dc690df33" exitCode=0 Oct 08 00:30:02 crc kubenswrapper[4967]: I1008 00:30:02.485643 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-25rkp" event={"ID":"5f8e39c1-cde6-4742-9d2a-cfbb47b5a14b","Type":"ContainerDied","Data":"56abe64eac56578021cf8baf3775ab7018104e2a15d62306943d047dc690df33"} Oct 08 00:30:02 crc kubenswrapper[4967]: I1008 00:30:02.723089 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331390-cv6kh" Oct 08 00:30:02 crc kubenswrapper[4967]: I1008 00:30:02.788689 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kskkw\" (UniqueName: \"kubernetes.io/projected/8c219dd4-8db4-4015-8df0-ab4989c5be71-kube-api-access-kskkw\") pod \"8c219dd4-8db4-4015-8df0-ab4989c5be71\" (UID: \"8c219dd4-8db4-4015-8df0-ab4989c5be71\") " Oct 08 00:30:02 crc kubenswrapper[4967]: I1008 00:30:02.788742 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8c219dd4-8db4-4015-8df0-ab4989c5be71-secret-volume\") pod \"8c219dd4-8db4-4015-8df0-ab4989c5be71\" (UID: \"8c219dd4-8db4-4015-8df0-ab4989c5be71\") " Oct 08 00:30:02 crc kubenswrapper[4967]: I1008 00:30:02.788785 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8c219dd4-8db4-4015-8df0-ab4989c5be71-config-volume\") pod \"8c219dd4-8db4-4015-8df0-ab4989c5be71\" (UID: \"8c219dd4-8db4-4015-8df0-ab4989c5be71\") " Oct 08 00:30:02 crc kubenswrapper[4967]: I1008 00:30:02.789653 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c219dd4-8db4-4015-8df0-ab4989c5be71-config-volume" (OuterVolumeSpecName: "config-volume") pod "8c219dd4-8db4-4015-8df0-ab4989c5be71" (UID: "8c219dd4-8db4-4015-8df0-ab4989c5be71"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:30:02 crc kubenswrapper[4967]: I1008 00:30:02.793745 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c219dd4-8db4-4015-8df0-ab4989c5be71-kube-api-access-kskkw" (OuterVolumeSpecName: "kube-api-access-kskkw") pod "8c219dd4-8db4-4015-8df0-ab4989c5be71" (UID: "8c219dd4-8db4-4015-8df0-ab4989c5be71"). InnerVolumeSpecName "kube-api-access-kskkw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:30:02 crc kubenswrapper[4967]: I1008 00:30:02.794007 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c219dd4-8db4-4015-8df0-ab4989c5be71-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "8c219dd4-8db4-4015-8df0-ab4989c5be71" (UID: "8c219dd4-8db4-4015-8df0-ab4989c5be71"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 00:30:02 crc kubenswrapper[4967]: I1008 00:30:02.890170 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kskkw\" (UniqueName: \"kubernetes.io/projected/8c219dd4-8db4-4015-8df0-ab4989c5be71-kube-api-access-kskkw\") on node \"crc\" DevicePath \"\"" Oct 08 00:30:02 crc kubenswrapper[4967]: I1008 00:30:02.890209 4967 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8c219dd4-8db4-4015-8df0-ab4989c5be71-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 08 00:30:02 crc kubenswrapper[4967]: I1008 00:30:02.890220 4967 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8c219dd4-8db4-4015-8df0-ab4989c5be71-config-volume\") on node \"crc\" DevicePath \"\"" Oct 08 00:30:03 crc kubenswrapper[4967]: I1008 00:30:03.214994 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/qdr-test"] Oct 08 00:30:03 crc kubenswrapper[4967]: E1008 00:30:03.215239 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c219dd4-8db4-4015-8df0-ab4989c5be71" containerName="collect-profiles" Oct 08 00:30:03 crc kubenswrapper[4967]: I1008 00:30:03.215254 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c219dd4-8db4-4015-8df0-ab4989c5be71" containerName="collect-profiles" Oct 08 00:30:03 crc kubenswrapper[4967]: I1008 00:30:03.215369 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c219dd4-8db4-4015-8df0-ab4989c5be71" containerName="collect-profiles" Oct 08 00:30:03 crc kubenswrapper[4967]: I1008 00:30:03.215774 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/qdr-test" Oct 08 00:30:03 crc kubenswrapper[4967]: I1008 00:30:03.220757 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"qdr-test-config" Oct 08 00:30:03 crc kubenswrapper[4967]: I1008 00:30:03.220999 4967 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-interconnect-selfsigned" Oct 08 00:30:03 crc kubenswrapper[4967]: I1008 00:30:03.226355 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/qdr-test"] Oct 08 00:30:03 crc kubenswrapper[4967]: I1008 00:30:03.296615 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"qdr-test-config\" (UniqueName: \"kubernetes.io/configmap/c8219ce4-173e-47ef-810d-6dc298346972-qdr-test-config\") pod \"qdr-test\" (UID: \"c8219ce4-173e-47ef-810d-6dc298346972\") " pod="service-telemetry/qdr-test" Oct 08 00:30:03 crc kubenswrapper[4967]: I1008 00:30:03.296706 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-selfsigned-cert\" (UniqueName: \"kubernetes.io/secret/c8219ce4-173e-47ef-810d-6dc298346972-default-interconnect-selfsigned-cert\") pod \"qdr-test\" (UID: \"c8219ce4-173e-47ef-810d-6dc298346972\") " pod="service-telemetry/qdr-test" Oct 08 00:30:03 crc kubenswrapper[4967]: I1008 00:30:03.296729 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mb5tn\" (UniqueName: \"kubernetes.io/projected/c8219ce4-173e-47ef-810d-6dc298346972-kube-api-access-mb5tn\") pod \"qdr-test\" (UID: \"c8219ce4-173e-47ef-810d-6dc298346972\") " pod="service-telemetry/qdr-test" Oct 08 00:30:03 crc kubenswrapper[4967]: I1008 00:30:03.398472 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"qdr-test-config\" (UniqueName: \"kubernetes.io/configmap/c8219ce4-173e-47ef-810d-6dc298346972-qdr-test-config\") pod \"qdr-test\" (UID: \"c8219ce4-173e-47ef-810d-6dc298346972\") " pod="service-telemetry/qdr-test" Oct 08 00:30:03 crc kubenswrapper[4967]: I1008 00:30:03.398535 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-interconnect-selfsigned-cert\" (UniqueName: \"kubernetes.io/secret/c8219ce4-173e-47ef-810d-6dc298346972-default-interconnect-selfsigned-cert\") pod \"qdr-test\" (UID: \"c8219ce4-173e-47ef-810d-6dc298346972\") " pod="service-telemetry/qdr-test" Oct 08 00:30:03 crc kubenswrapper[4967]: I1008 00:30:03.398555 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mb5tn\" (UniqueName: \"kubernetes.io/projected/c8219ce4-173e-47ef-810d-6dc298346972-kube-api-access-mb5tn\") pod \"qdr-test\" (UID: \"c8219ce4-173e-47ef-810d-6dc298346972\") " pod="service-telemetry/qdr-test" Oct 08 00:30:03 crc kubenswrapper[4967]: I1008 00:30:03.399249 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"qdr-test-config\" (UniqueName: \"kubernetes.io/configmap/c8219ce4-173e-47ef-810d-6dc298346972-qdr-test-config\") pod \"qdr-test\" (UID: \"c8219ce4-173e-47ef-810d-6dc298346972\") " pod="service-telemetry/qdr-test" Oct 08 00:30:03 crc kubenswrapper[4967]: I1008 00:30:03.406620 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-interconnect-selfsigned-cert\" (UniqueName: \"kubernetes.io/secret/c8219ce4-173e-47ef-810d-6dc298346972-default-interconnect-selfsigned-cert\") pod \"qdr-test\" (UID: \"c8219ce4-173e-47ef-810d-6dc298346972\") " pod="service-telemetry/qdr-test" Oct 08 00:30:03 crc kubenswrapper[4967]: I1008 00:30:03.413482 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mb5tn\" (UniqueName: \"kubernetes.io/projected/c8219ce4-173e-47ef-810d-6dc298346972-kube-api-access-mb5tn\") pod \"qdr-test\" (UID: \"c8219ce4-173e-47ef-810d-6dc298346972\") " pod="service-telemetry/qdr-test" Oct 08 00:30:03 crc kubenswrapper[4967]: I1008 00:30:03.520548 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29331390-cv6kh" Oct 08 00:30:03 crc kubenswrapper[4967]: I1008 00:30:03.520529 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29331390-cv6kh" event={"ID":"8c219dd4-8db4-4015-8df0-ab4989c5be71","Type":"ContainerDied","Data":"ca71e630d09e321aff646c4edbbd124bcdfc8aab41d137c358f38348b6434124"} Oct 08 00:30:03 crc kubenswrapper[4967]: I1008 00:30:03.521050 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ca71e630d09e321aff646c4edbbd124bcdfc8aab41d137c358f38348b6434124" Oct 08 00:30:03 crc kubenswrapper[4967]: I1008 00:30:03.525307 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-25rkp" event={"ID":"5f8e39c1-cde6-4742-9d2a-cfbb47b5a14b","Type":"ContainerStarted","Data":"f18d34e363d26ead48492bf08acd050701fc7a6377aca360f325434600dfc07d"} Oct 08 00:30:03 crc kubenswrapper[4967]: I1008 00:30:03.545983 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-25rkp" podStartSLOduration=2.8517822539999997 podStartE2EDuration="5.545961563s" podCreationTimestamp="2025-10-08 00:29:58 +0000 UTC" firstStartedPulling="2025-10-08 00:30:00.467061369 +0000 UTC m=+1399.117596058" lastFinishedPulling="2025-10-08 00:30:03.161240718 +0000 UTC m=+1401.811775367" observedRunningTime="2025-10-08 00:30:03.539775487 +0000 UTC m=+1402.190310126" watchObservedRunningTime="2025-10-08 00:30:03.545961563 +0000 UTC m=+1402.196496222" Oct 08 00:30:03 crc kubenswrapper[4967]: I1008 00:30:03.601093 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/qdr-test" Oct 08 00:30:04 crc kubenswrapper[4967]: I1008 00:30:04.018225 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/qdr-test"] Oct 08 00:30:04 crc kubenswrapper[4967]: W1008 00:30:04.037270 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc8219ce4_173e_47ef_810d_6dc298346972.slice/crio-9d180ac7919f4086bbd66c97aac8590fe0e62d831eb4d40c1fbeeed49908d830 WatchSource:0}: Error finding container 9d180ac7919f4086bbd66c97aac8590fe0e62d831eb4d40c1fbeeed49908d830: Status 404 returned error can't find the container with id 9d180ac7919f4086bbd66c97aac8590fe0e62d831eb4d40c1fbeeed49908d830 Oct 08 00:30:04 crc kubenswrapper[4967]: I1008 00:30:04.517206 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-txtvn"] Oct 08 00:30:04 crc kubenswrapper[4967]: I1008 00:30:04.518453 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-txtvn" Oct 08 00:30:04 crc kubenswrapper[4967]: I1008 00:30:04.529259 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-txtvn"] Oct 08 00:30:04 crc kubenswrapper[4967]: I1008 00:30:04.533688 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/qdr-test" event={"ID":"c8219ce4-173e-47ef-810d-6dc298346972","Type":"ContainerStarted","Data":"9d180ac7919f4086bbd66c97aac8590fe0e62d831eb4d40c1fbeeed49908d830"} Oct 08 00:30:04 crc kubenswrapper[4967]: I1008 00:30:04.613159 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b156a9a-950b-40da-b51e-e9774ce7b027-catalog-content\") pod \"community-operators-txtvn\" (UID: \"6b156a9a-950b-40da-b51e-e9774ce7b027\") " pod="openshift-marketplace/community-operators-txtvn" Oct 08 00:30:04 crc kubenswrapper[4967]: I1008 00:30:04.613218 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hzrbl\" (UniqueName: \"kubernetes.io/projected/6b156a9a-950b-40da-b51e-e9774ce7b027-kube-api-access-hzrbl\") pod \"community-operators-txtvn\" (UID: \"6b156a9a-950b-40da-b51e-e9774ce7b027\") " pod="openshift-marketplace/community-operators-txtvn" Oct 08 00:30:04 crc kubenswrapper[4967]: I1008 00:30:04.613250 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b156a9a-950b-40da-b51e-e9774ce7b027-utilities\") pod \"community-operators-txtvn\" (UID: \"6b156a9a-950b-40da-b51e-e9774ce7b027\") " pod="openshift-marketplace/community-operators-txtvn" Oct 08 00:30:04 crc kubenswrapper[4967]: I1008 00:30:04.714326 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hzrbl\" (UniqueName: \"kubernetes.io/projected/6b156a9a-950b-40da-b51e-e9774ce7b027-kube-api-access-hzrbl\") pod \"community-operators-txtvn\" (UID: \"6b156a9a-950b-40da-b51e-e9774ce7b027\") " pod="openshift-marketplace/community-operators-txtvn" Oct 08 00:30:04 crc kubenswrapper[4967]: I1008 00:30:04.714381 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b156a9a-950b-40da-b51e-e9774ce7b027-utilities\") pod \"community-operators-txtvn\" (UID: \"6b156a9a-950b-40da-b51e-e9774ce7b027\") " pod="openshift-marketplace/community-operators-txtvn" Oct 08 00:30:04 crc kubenswrapper[4967]: I1008 00:30:04.714462 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b156a9a-950b-40da-b51e-e9774ce7b027-catalog-content\") pod \"community-operators-txtvn\" (UID: \"6b156a9a-950b-40da-b51e-e9774ce7b027\") " pod="openshift-marketplace/community-operators-txtvn" Oct 08 00:30:04 crc kubenswrapper[4967]: I1008 00:30:04.715003 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b156a9a-950b-40da-b51e-e9774ce7b027-catalog-content\") pod \"community-operators-txtvn\" (UID: \"6b156a9a-950b-40da-b51e-e9774ce7b027\") " pod="openshift-marketplace/community-operators-txtvn" Oct 08 00:30:04 crc kubenswrapper[4967]: I1008 00:30:04.715155 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b156a9a-950b-40da-b51e-e9774ce7b027-utilities\") pod \"community-operators-txtvn\" (UID: \"6b156a9a-950b-40da-b51e-e9774ce7b027\") " pod="openshift-marketplace/community-operators-txtvn" Oct 08 00:30:04 crc kubenswrapper[4967]: I1008 00:30:04.734435 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hzrbl\" (UniqueName: \"kubernetes.io/projected/6b156a9a-950b-40da-b51e-e9774ce7b027-kube-api-access-hzrbl\") pod \"community-operators-txtvn\" (UID: \"6b156a9a-950b-40da-b51e-e9774ce7b027\") " pod="openshift-marketplace/community-operators-txtvn" Oct 08 00:30:04 crc kubenswrapper[4967]: I1008 00:30:04.836561 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-txtvn" Oct 08 00:30:05 crc kubenswrapper[4967]: I1008 00:30:05.348349 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-txtvn"] Oct 08 00:30:05 crc kubenswrapper[4967]: I1008 00:30:05.542609 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-txtvn" event={"ID":"6b156a9a-950b-40da-b51e-e9774ce7b027","Type":"ContainerStarted","Data":"8a8da8f4c2dca8a8072b541f1eae6d865af9e930337fe713b23099c5fd630671"} Oct 08 00:30:06 crc kubenswrapper[4967]: I1008 00:30:06.551876 4967 generic.go:334] "Generic (PLEG): container finished" podID="6b156a9a-950b-40da-b51e-e9774ce7b027" containerID="6658ff11c323ece9282af45a9d61f28182be1130e3e13a92e006b90e43287621" exitCode=0 Oct 08 00:30:06 crc kubenswrapper[4967]: I1008 00:30:06.551917 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-txtvn" event={"ID":"6b156a9a-950b-40da-b51e-e9774ce7b027","Type":"ContainerDied","Data":"6658ff11c323ece9282af45a9d61f28182be1130e3e13a92e006b90e43287621"} Oct 08 00:30:06 crc kubenswrapper[4967]: I1008 00:30:06.920450 4967 patch_prober.go:28] interesting pod/machine-config-daemon-xvdtp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 00:30:06 crc kubenswrapper[4967]: I1008 00:30:06.920847 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" podUID="9461162a-c27d-465e-9ca7-2d3936877cd8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 00:30:08 crc kubenswrapper[4967]: I1008 00:30:08.570062 4967 generic.go:334] "Generic (PLEG): container finished" podID="6b156a9a-950b-40da-b51e-e9774ce7b027" containerID="589d45040ff6bcdc265de3e03ee5be21ef82d1cd3ce65a30de12e71ca59ba3f7" exitCode=0 Oct 08 00:30:08 crc kubenswrapper[4967]: I1008 00:30:08.570120 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-txtvn" event={"ID":"6b156a9a-950b-40da-b51e-e9774ce7b027","Type":"ContainerDied","Data":"589d45040ff6bcdc265de3e03ee5be21ef82d1cd3ce65a30de12e71ca59ba3f7"} Oct 08 00:30:09 crc kubenswrapper[4967]: I1008 00:30:09.067427 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-25rkp" Oct 08 00:30:09 crc kubenswrapper[4967]: I1008 00:30:09.068441 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-25rkp" Oct 08 00:30:09 crc kubenswrapper[4967]: I1008 00:30:09.131654 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-25rkp" Oct 08 00:30:09 crc kubenswrapper[4967]: I1008 00:30:09.628974 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-25rkp" Oct 08 00:30:10 crc kubenswrapper[4967]: I1008 00:30:10.119070 4967 scope.go:117] "RemoveContainer" containerID="172a90c9893e70774c41d3027c26eac57aca1df3d0c370d72d1a1e7bd879cb82" Oct 08 00:30:11 crc kubenswrapper[4967]: I1008 00:30:11.118439 4967 scope.go:117] "RemoveContainer" containerID="71734357db2a83fc820ed005863935126520931ebbe4df8c40ea34276b7831f9" Oct 08 00:30:11 crc kubenswrapper[4967]: I1008 00:30:11.319141 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-25rkp"] Oct 08 00:30:11 crc kubenswrapper[4967]: I1008 00:30:11.528491 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-jk962"] Oct 08 00:30:11 crc kubenswrapper[4967]: I1008 00:30:11.532922 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jk962" Oct 08 00:30:11 crc kubenswrapper[4967]: I1008 00:30:11.545048 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jk962"] Oct 08 00:30:11 crc kubenswrapper[4967]: I1008 00:30:11.630594 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd3320fb-5873-4815-ab64-c7fe697c0532-utilities\") pod \"certified-operators-jk962\" (UID: \"dd3320fb-5873-4815-ab64-c7fe697c0532\") " pod="openshift-marketplace/certified-operators-jk962" Oct 08 00:30:11 crc kubenswrapper[4967]: I1008 00:30:11.630697 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqqbd\" (UniqueName: \"kubernetes.io/projected/dd3320fb-5873-4815-ab64-c7fe697c0532-kube-api-access-dqqbd\") pod \"certified-operators-jk962\" (UID: \"dd3320fb-5873-4815-ab64-c7fe697c0532\") " pod="openshift-marketplace/certified-operators-jk962" Oct 08 00:30:11 crc kubenswrapper[4967]: I1008 00:30:11.630724 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd3320fb-5873-4815-ab64-c7fe697c0532-catalog-content\") pod \"certified-operators-jk962\" (UID: \"dd3320fb-5873-4815-ab64-c7fe697c0532\") " pod="openshift-marketplace/certified-operators-jk962" Oct 08 00:30:11 crc kubenswrapper[4967]: I1008 00:30:11.732191 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dqqbd\" (UniqueName: \"kubernetes.io/projected/dd3320fb-5873-4815-ab64-c7fe697c0532-kube-api-access-dqqbd\") pod \"certified-operators-jk962\" (UID: \"dd3320fb-5873-4815-ab64-c7fe697c0532\") " pod="openshift-marketplace/certified-operators-jk962" Oct 08 00:30:11 crc kubenswrapper[4967]: I1008 00:30:11.732256 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd3320fb-5873-4815-ab64-c7fe697c0532-catalog-content\") pod \"certified-operators-jk962\" (UID: \"dd3320fb-5873-4815-ab64-c7fe697c0532\") " pod="openshift-marketplace/certified-operators-jk962" Oct 08 00:30:11 crc kubenswrapper[4967]: I1008 00:30:11.732316 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd3320fb-5873-4815-ab64-c7fe697c0532-utilities\") pod \"certified-operators-jk962\" (UID: \"dd3320fb-5873-4815-ab64-c7fe697c0532\") " pod="openshift-marketplace/certified-operators-jk962" Oct 08 00:30:11 crc kubenswrapper[4967]: I1008 00:30:11.732801 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd3320fb-5873-4815-ab64-c7fe697c0532-catalog-content\") pod \"certified-operators-jk962\" (UID: \"dd3320fb-5873-4815-ab64-c7fe697c0532\") " pod="openshift-marketplace/certified-operators-jk962" Oct 08 00:30:11 crc kubenswrapper[4967]: I1008 00:30:11.732833 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd3320fb-5873-4815-ab64-c7fe697c0532-utilities\") pod \"certified-operators-jk962\" (UID: \"dd3320fb-5873-4815-ab64-c7fe697c0532\") " pod="openshift-marketplace/certified-operators-jk962" Oct 08 00:30:11 crc kubenswrapper[4967]: I1008 00:30:11.784051 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dqqbd\" (UniqueName: \"kubernetes.io/projected/dd3320fb-5873-4815-ab64-c7fe697c0532-kube-api-access-dqqbd\") pod \"certified-operators-jk962\" (UID: \"dd3320fb-5873-4815-ab64-c7fe697c0532\") " pod="openshift-marketplace/certified-operators-jk962" Oct 08 00:30:11 crc kubenswrapper[4967]: I1008 00:30:11.877426 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jk962" Oct 08 00:30:12 crc kubenswrapper[4967]: I1008 00:30:12.615209 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-25rkp" podUID="5f8e39c1-cde6-4742-9d2a-cfbb47b5a14b" containerName="registry-server" containerID="cri-o://f18d34e363d26ead48492bf08acd050701fc7a6377aca360f325434600dfc07d" gracePeriod=2 Oct 08 00:30:13 crc kubenswrapper[4967]: I1008 00:30:13.118171 4967 scope.go:117] "RemoveContainer" containerID="45616f6aeab51b4f086393063c6d90a2738cdb3308a9acd32ac19dfa9003472d" Oct 08 00:30:13 crc kubenswrapper[4967]: I1008 00:30:13.118327 4967 scope.go:117] "RemoveContainer" containerID="5708a0875e9b4587f5e7e31ae66f8fa5ab5140fa37e4d1c8f1c2a9aceef79143" Oct 08 00:30:13 crc kubenswrapper[4967]: I1008 00:30:13.644646 4967 generic.go:334] "Generic (PLEG): container finished" podID="5f8e39c1-cde6-4742-9d2a-cfbb47b5a14b" containerID="f18d34e363d26ead48492bf08acd050701fc7a6377aca360f325434600dfc07d" exitCode=0 Oct 08 00:30:13 crc kubenswrapper[4967]: I1008 00:30:13.644709 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-25rkp" event={"ID":"5f8e39c1-cde6-4742-9d2a-cfbb47b5a14b","Type":"ContainerDied","Data":"f18d34e363d26ead48492bf08acd050701fc7a6377aca360f325434600dfc07d"} Oct 08 00:30:14 crc kubenswrapper[4967]: I1008 00:30:14.119110 4967 scope.go:117] "RemoveContainer" containerID="b7d19b7c1b6d22b5b39b4ecccaa58ab4f49169900f0d047f09cab046927a8037" Oct 08 00:30:14 crc kubenswrapper[4967]: I1008 00:30:14.121711 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-25rkp" Oct 08 00:30:14 crc kubenswrapper[4967]: I1008 00:30:14.194063 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f8e39c1-cde6-4742-9d2a-cfbb47b5a14b-utilities\") pod \"5f8e39c1-cde6-4742-9d2a-cfbb47b5a14b\" (UID: \"5f8e39c1-cde6-4742-9d2a-cfbb47b5a14b\") " Oct 08 00:30:14 crc kubenswrapper[4967]: I1008 00:30:14.194148 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cbjl5\" (UniqueName: \"kubernetes.io/projected/5f8e39c1-cde6-4742-9d2a-cfbb47b5a14b-kube-api-access-cbjl5\") pod \"5f8e39c1-cde6-4742-9d2a-cfbb47b5a14b\" (UID: \"5f8e39c1-cde6-4742-9d2a-cfbb47b5a14b\") " Oct 08 00:30:14 crc kubenswrapper[4967]: I1008 00:30:14.194166 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f8e39c1-cde6-4742-9d2a-cfbb47b5a14b-catalog-content\") pod \"5f8e39c1-cde6-4742-9d2a-cfbb47b5a14b\" (UID: \"5f8e39c1-cde6-4742-9d2a-cfbb47b5a14b\") " Oct 08 00:30:14 crc kubenswrapper[4967]: I1008 00:30:14.195488 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5f8e39c1-cde6-4742-9d2a-cfbb47b5a14b-utilities" (OuterVolumeSpecName: "utilities") pod "5f8e39c1-cde6-4742-9d2a-cfbb47b5a14b" (UID: "5f8e39c1-cde6-4742-9d2a-cfbb47b5a14b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:30:14 crc kubenswrapper[4967]: I1008 00:30:14.213202 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f8e39c1-cde6-4742-9d2a-cfbb47b5a14b-kube-api-access-cbjl5" (OuterVolumeSpecName: "kube-api-access-cbjl5") pod "5f8e39c1-cde6-4742-9d2a-cfbb47b5a14b" (UID: "5f8e39c1-cde6-4742-9d2a-cfbb47b5a14b"). InnerVolumeSpecName "kube-api-access-cbjl5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:30:14 crc kubenswrapper[4967]: I1008 00:30:14.295649 4967 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f8e39c1-cde6-4742-9d2a-cfbb47b5a14b-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 00:30:14 crc kubenswrapper[4967]: I1008 00:30:14.295703 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cbjl5\" (UniqueName: \"kubernetes.io/projected/5f8e39c1-cde6-4742-9d2a-cfbb47b5a14b-kube-api-access-cbjl5\") on node \"crc\" DevicePath \"\"" Oct 08 00:30:14 crc kubenswrapper[4967]: I1008 00:30:14.307923 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5f8e39c1-cde6-4742-9d2a-cfbb47b5a14b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5f8e39c1-cde6-4742-9d2a-cfbb47b5a14b" (UID: "5f8e39c1-cde6-4742-9d2a-cfbb47b5a14b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:30:14 crc kubenswrapper[4967]: I1008 00:30:14.399121 4967 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f8e39c1-cde6-4742-9d2a-cfbb47b5a14b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 00:30:14 crc kubenswrapper[4967]: I1008 00:30:14.514801 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jk962"] Oct 08 00:30:14 crc kubenswrapper[4967]: W1008 00:30:14.519628 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddd3320fb_5873_4815_ab64_c7fe697c0532.slice/crio-c872194e3b5ad502f28189aafe5272036fad0ebb02478f09c37d0c4921b92514 WatchSource:0}: Error finding container c872194e3b5ad502f28189aafe5272036fad0ebb02478f09c37d0c4921b92514: Status 404 returned error can't find the container with id c872194e3b5ad502f28189aafe5272036fad0ebb02478f09c37d0c4921b92514 Oct 08 00:30:14 crc kubenswrapper[4967]: I1008 00:30:14.655080 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-57948895dc-dlql4" event={"ID":"1f1c83bb-ff16-45ff-8fe5-e1b7e9d5166b","Type":"ContainerStarted","Data":"05ba4995a912475e24bd8c6b04abf02799309117a43fbd37d4b4d5f54820583f"} Oct 08 00:30:14 crc kubenswrapper[4967]: I1008 00:30:14.657660 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-5d47c9955c-t24dp" event={"ID":"b3e181c2-79f9-47be-b4fd-f9f05599395f","Type":"ContainerStarted","Data":"4407309ed28ab3aab57f73251c45dfb7c490c587f960b68c1bc64882d02068d4"} Oct 08 00:30:14 crc kubenswrapper[4967]: I1008 00:30:14.659507 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-event-smartgateway-6f7464cc9d-bjhng" event={"ID":"84dadc03-b18c-490c-adb2-9d3f605865d1","Type":"ContainerStarted","Data":"b9b6f2fbf81a04ae6defed264aeb84e3e2197e81193587b349c1c8169258ab75"} Oct 08 00:30:14 crc kubenswrapper[4967]: I1008 00:30:14.661463 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-5759b4d97-rwnt8" event={"ID":"a9da825c-48d2-477d-b8c5-1ac878239c37","Type":"ContainerStarted","Data":"9aa266ae834d35f065531da1594fa84ad9f2a97b0fe5d3bebe7306d1f6ff503b"} Oct 08 00:30:14 crc kubenswrapper[4967]: I1008 00:30:14.662845 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/qdr-test" event={"ID":"c8219ce4-173e-47ef-810d-6dc298346972","Type":"ContainerStarted","Data":"3eec603a34ad4aed01ce1e7942d19855874cede3211a5d19cf4fca6dccb1bf79"} Oct 08 00:30:14 crc kubenswrapper[4967]: I1008 00:30:14.664892 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jk962" event={"ID":"dd3320fb-5873-4815-ab64-c7fe697c0532","Type":"ContainerStarted","Data":"b25422ee76ce98a45191d83b9fedcea09379d97565986640b5117b3e58cc1aff"} Oct 08 00:30:14 crc kubenswrapper[4967]: I1008 00:30:14.664923 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jk962" event={"ID":"dd3320fb-5873-4815-ab64-c7fe697c0532","Type":"ContainerStarted","Data":"c872194e3b5ad502f28189aafe5272036fad0ebb02478f09c37d0c4921b92514"} Oct 08 00:30:14 crc kubenswrapper[4967]: I1008 00:30:14.667951 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-7cd87f9766-qcv8n" event={"ID":"ddba1803-397c-4daf-bc56-4857967229d1","Type":"ContainerStarted","Data":"c874355a890c8b8d9204b65202953b760c1c4a770fd49d586e3e93e6e34c1dba"} Oct 08 00:30:14 crc kubenswrapper[4967]: I1008 00:30:14.671084 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-25rkp" event={"ID":"5f8e39c1-cde6-4742-9d2a-cfbb47b5a14b","Type":"ContainerDied","Data":"e0a8dca80f5e3eb594da12a03ca70d682ec48a86422c9c9bf19a6126a8dc81c6"} Oct 08 00:30:14 crc kubenswrapper[4967]: I1008 00:30:14.671122 4967 scope.go:117] "RemoveContainer" containerID="f18d34e363d26ead48492bf08acd050701fc7a6377aca360f325434600dfc07d" Oct 08 00:30:14 crc kubenswrapper[4967]: I1008 00:30:14.671221 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-25rkp" Oct 08 00:30:14 crc kubenswrapper[4967]: I1008 00:30:14.682014 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-txtvn" event={"ID":"6b156a9a-950b-40da-b51e-e9774ce7b027","Type":"ContainerStarted","Data":"fe2fdead628193cfbf79bf235241329ed1ed7b11cf34ad489a04a0ebee862660"} Oct 08 00:30:14 crc kubenswrapper[4967]: I1008 00:30:14.727933 4967 scope.go:117] "RemoveContainer" containerID="56abe64eac56578021cf8baf3775ab7018104e2a15d62306943d047dc690df33" Oct 08 00:30:14 crc kubenswrapper[4967]: I1008 00:30:14.791789 4967 scope.go:117] "RemoveContainer" containerID="ba88b4442e2df98ac6775b84a3453e241f2e9443a9cc6b61d2442ab80a2cc30c" Oct 08 00:30:14 crc kubenswrapper[4967]: I1008 00:30:14.808308 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/qdr-test" podStartSLOduration=1.661095284 podStartE2EDuration="11.80828499s" podCreationTimestamp="2025-10-08 00:30:03 +0000 UTC" firstStartedPulling="2025-10-08 00:30:04.038905949 +0000 UTC m=+1402.689440598" lastFinishedPulling="2025-10-08 00:30:14.186095655 +0000 UTC m=+1412.836630304" observedRunningTime="2025-10-08 00:30:14.807001118 +0000 UTC m=+1413.457535777" watchObservedRunningTime="2025-10-08 00:30:14.80828499 +0000 UTC m=+1413.458819639" Oct 08 00:30:14 crc kubenswrapper[4967]: I1008 00:30:14.838129 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-txtvn" Oct 08 00:30:14 crc kubenswrapper[4967]: I1008 00:30:14.838160 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-txtvn" Oct 08 00:30:14 crc kubenswrapper[4967]: I1008 00:30:14.861239 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-txtvn" podStartSLOduration=3.311139374 podStartE2EDuration="10.861219572s" podCreationTimestamp="2025-10-08 00:30:04 +0000 UTC" firstStartedPulling="2025-10-08 00:30:06.554139417 +0000 UTC m=+1405.204674066" lastFinishedPulling="2025-10-08 00:30:14.104219615 +0000 UTC m=+1412.754754264" observedRunningTime="2025-10-08 00:30:14.858768 +0000 UTC m=+1413.509302669" watchObservedRunningTime="2025-10-08 00:30:14.861219572 +0000 UTC m=+1413.511754221" Oct 08 00:30:14 crc kubenswrapper[4967]: I1008 00:30:14.883855 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-25rkp"] Oct 08 00:30:14 crc kubenswrapper[4967]: I1008 00:30:14.887897 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-25rkp"] Oct 08 00:30:15 crc kubenswrapper[4967]: I1008 00:30:15.130533 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/stf-smoketest-smoke1-m46pn"] Oct 08 00:30:15 crc kubenswrapper[4967]: E1008 00:30:15.130818 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f8e39c1-cde6-4742-9d2a-cfbb47b5a14b" containerName="extract-content" Oct 08 00:30:15 crc kubenswrapper[4967]: I1008 00:30:15.130833 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f8e39c1-cde6-4742-9d2a-cfbb47b5a14b" containerName="extract-content" Oct 08 00:30:15 crc kubenswrapper[4967]: E1008 00:30:15.130856 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f8e39c1-cde6-4742-9d2a-cfbb47b5a14b" containerName="registry-server" Oct 08 00:30:15 crc kubenswrapper[4967]: I1008 00:30:15.130862 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f8e39c1-cde6-4742-9d2a-cfbb47b5a14b" containerName="registry-server" Oct 08 00:30:15 crc kubenswrapper[4967]: E1008 00:30:15.130872 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f8e39c1-cde6-4742-9d2a-cfbb47b5a14b" containerName="extract-utilities" Oct 08 00:30:15 crc kubenswrapper[4967]: I1008 00:30:15.130877 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f8e39c1-cde6-4742-9d2a-cfbb47b5a14b" containerName="extract-utilities" Oct 08 00:30:15 crc kubenswrapper[4967]: I1008 00:30:15.130988 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f8e39c1-cde6-4742-9d2a-cfbb47b5a14b" containerName="registry-server" Oct 08 00:30:15 crc kubenswrapper[4967]: I1008 00:30:15.131594 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/stf-smoketest-smoke1-m46pn" Oct 08 00:30:15 crc kubenswrapper[4967]: I1008 00:30:15.135329 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"stf-smoketest-collectd-entrypoint-script" Oct 08 00:30:15 crc kubenswrapper[4967]: I1008 00:30:15.135726 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"stf-smoketest-collectd-config" Oct 08 00:30:15 crc kubenswrapper[4967]: I1008 00:30:15.136044 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"stf-smoketest-ceilometer-entrypoint-script" Oct 08 00:30:15 crc kubenswrapper[4967]: I1008 00:30:15.136328 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"stf-smoketest-ceilometer-publisher" Oct 08 00:30:15 crc kubenswrapper[4967]: I1008 00:30:15.137112 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"stf-smoketest-sensubility-config" Oct 08 00:30:15 crc kubenswrapper[4967]: I1008 00:30:15.145478 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"stf-smoketest-healthcheck-log" Oct 08 00:30:15 crc kubenswrapper[4967]: I1008 00:30:15.146223 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/stf-smoketest-smoke1-m46pn"] Oct 08 00:30:15 crc kubenswrapper[4967]: I1008 00:30:15.213705 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gq8mp\" (UniqueName: \"kubernetes.io/projected/69b88376-a0bc-4864-bdbb-a445ca0c90b4-kube-api-access-gq8mp\") pod \"stf-smoketest-smoke1-m46pn\" (UID: \"69b88376-a0bc-4864-bdbb-a445ca0c90b4\") " pod="service-telemetry/stf-smoketest-smoke1-m46pn" Oct 08 00:30:15 crc kubenswrapper[4967]: I1008 00:30:15.213766 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collectd-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/69b88376-a0bc-4864-bdbb-a445ca0c90b4-collectd-entrypoint-script\") pod \"stf-smoketest-smoke1-m46pn\" (UID: \"69b88376-a0bc-4864-bdbb-a445ca0c90b4\") " pod="service-telemetry/stf-smoketest-smoke1-m46pn" Oct 08 00:30:15 crc kubenswrapper[4967]: I1008 00:30:15.213830 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sensubility-config\" (UniqueName: \"kubernetes.io/configmap/69b88376-a0bc-4864-bdbb-a445ca0c90b4-sensubility-config\") pod \"stf-smoketest-smoke1-m46pn\" (UID: \"69b88376-a0bc-4864-bdbb-a445ca0c90b4\") " pod="service-telemetry/stf-smoketest-smoke1-m46pn" Oct 08 00:30:15 crc kubenswrapper[4967]: I1008 00:30:15.213957 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collectd-config\" (UniqueName: \"kubernetes.io/configmap/69b88376-a0bc-4864-bdbb-a445ca0c90b4-collectd-config\") pod \"stf-smoketest-smoke1-m46pn\" (UID: \"69b88376-a0bc-4864-bdbb-a445ca0c90b4\") " pod="service-telemetry/stf-smoketest-smoke1-m46pn" Oct 08 00:30:15 crc kubenswrapper[4967]: I1008 00:30:15.213981 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"healthcheck-log\" (UniqueName: \"kubernetes.io/configmap/69b88376-a0bc-4864-bdbb-a445ca0c90b4-healthcheck-log\") pod \"stf-smoketest-smoke1-m46pn\" (UID: \"69b88376-a0bc-4864-bdbb-a445ca0c90b4\") " pod="service-telemetry/stf-smoketest-smoke1-m46pn" Oct 08 00:30:15 crc kubenswrapper[4967]: I1008 00:30:15.214061 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/69b88376-a0bc-4864-bdbb-a445ca0c90b4-ceilometer-entrypoint-script\") pod \"stf-smoketest-smoke1-m46pn\" (UID: \"69b88376-a0bc-4864-bdbb-a445ca0c90b4\") " pod="service-telemetry/stf-smoketest-smoke1-m46pn" Oct 08 00:30:15 crc kubenswrapper[4967]: I1008 00:30:15.214125 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-publisher\" (UniqueName: \"kubernetes.io/configmap/69b88376-a0bc-4864-bdbb-a445ca0c90b4-ceilometer-publisher\") pod \"stf-smoketest-smoke1-m46pn\" (UID: \"69b88376-a0bc-4864-bdbb-a445ca0c90b4\") " pod="service-telemetry/stf-smoketest-smoke1-m46pn" Oct 08 00:30:15 crc kubenswrapper[4967]: I1008 00:30:15.315952 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gq8mp\" (UniqueName: \"kubernetes.io/projected/69b88376-a0bc-4864-bdbb-a445ca0c90b4-kube-api-access-gq8mp\") pod \"stf-smoketest-smoke1-m46pn\" (UID: \"69b88376-a0bc-4864-bdbb-a445ca0c90b4\") " pod="service-telemetry/stf-smoketest-smoke1-m46pn" Oct 08 00:30:15 crc kubenswrapper[4967]: I1008 00:30:15.316012 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collectd-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/69b88376-a0bc-4864-bdbb-a445ca0c90b4-collectd-entrypoint-script\") pod \"stf-smoketest-smoke1-m46pn\" (UID: \"69b88376-a0bc-4864-bdbb-a445ca0c90b4\") " pod="service-telemetry/stf-smoketest-smoke1-m46pn" Oct 08 00:30:15 crc kubenswrapper[4967]: I1008 00:30:15.316043 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sensubility-config\" (UniqueName: \"kubernetes.io/configmap/69b88376-a0bc-4864-bdbb-a445ca0c90b4-sensubility-config\") pod \"stf-smoketest-smoke1-m46pn\" (UID: \"69b88376-a0bc-4864-bdbb-a445ca0c90b4\") " pod="service-telemetry/stf-smoketest-smoke1-m46pn" Oct 08 00:30:15 crc kubenswrapper[4967]: I1008 00:30:15.316098 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collectd-config\" (UniqueName: \"kubernetes.io/configmap/69b88376-a0bc-4864-bdbb-a445ca0c90b4-collectd-config\") pod \"stf-smoketest-smoke1-m46pn\" (UID: \"69b88376-a0bc-4864-bdbb-a445ca0c90b4\") " pod="service-telemetry/stf-smoketest-smoke1-m46pn" Oct 08 00:30:15 crc kubenswrapper[4967]: I1008 00:30:15.316117 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"healthcheck-log\" (UniqueName: \"kubernetes.io/configmap/69b88376-a0bc-4864-bdbb-a445ca0c90b4-healthcheck-log\") pod \"stf-smoketest-smoke1-m46pn\" (UID: \"69b88376-a0bc-4864-bdbb-a445ca0c90b4\") " pod="service-telemetry/stf-smoketest-smoke1-m46pn" Oct 08 00:30:15 crc kubenswrapper[4967]: I1008 00:30:15.317175 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collectd-config\" (UniqueName: \"kubernetes.io/configmap/69b88376-a0bc-4864-bdbb-a445ca0c90b4-collectd-config\") pod \"stf-smoketest-smoke1-m46pn\" (UID: \"69b88376-a0bc-4864-bdbb-a445ca0c90b4\") " pod="service-telemetry/stf-smoketest-smoke1-m46pn" Oct 08 00:30:15 crc kubenswrapper[4967]: I1008 00:30:15.317192 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sensubility-config\" (UniqueName: \"kubernetes.io/configmap/69b88376-a0bc-4864-bdbb-a445ca0c90b4-sensubility-config\") pod \"stf-smoketest-smoke1-m46pn\" (UID: \"69b88376-a0bc-4864-bdbb-a445ca0c90b4\") " pod="service-telemetry/stf-smoketest-smoke1-m46pn" Oct 08 00:30:15 crc kubenswrapper[4967]: I1008 00:30:15.317252 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/69b88376-a0bc-4864-bdbb-a445ca0c90b4-ceilometer-entrypoint-script\") pod \"stf-smoketest-smoke1-m46pn\" (UID: \"69b88376-a0bc-4864-bdbb-a445ca0c90b4\") " pod="service-telemetry/stf-smoketest-smoke1-m46pn" Oct 08 00:30:15 crc kubenswrapper[4967]: I1008 00:30:15.317278 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collectd-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/69b88376-a0bc-4864-bdbb-a445ca0c90b4-collectd-entrypoint-script\") pod \"stf-smoketest-smoke1-m46pn\" (UID: \"69b88376-a0bc-4864-bdbb-a445ca0c90b4\") " pod="service-telemetry/stf-smoketest-smoke1-m46pn" Oct 08 00:30:15 crc kubenswrapper[4967]: I1008 00:30:15.317323 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"healthcheck-log\" (UniqueName: \"kubernetes.io/configmap/69b88376-a0bc-4864-bdbb-a445ca0c90b4-healthcheck-log\") pod \"stf-smoketest-smoke1-m46pn\" (UID: \"69b88376-a0bc-4864-bdbb-a445ca0c90b4\") " pod="service-telemetry/stf-smoketest-smoke1-m46pn" Oct 08 00:30:15 crc kubenswrapper[4967]: I1008 00:30:15.317335 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-publisher\" (UniqueName: \"kubernetes.io/configmap/69b88376-a0bc-4864-bdbb-a445ca0c90b4-ceilometer-publisher\") pod \"stf-smoketest-smoke1-m46pn\" (UID: \"69b88376-a0bc-4864-bdbb-a445ca0c90b4\") " pod="service-telemetry/stf-smoketest-smoke1-m46pn" Oct 08 00:30:15 crc kubenswrapper[4967]: I1008 00:30:15.318034 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-publisher\" (UniqueName: \"kubernetes.io/configmap/69b88376-a0bc-4864-bdbb-a445ca0c90b4-ceilometer-publisher\") pod \"stf-smoketest-smoke1-m46pn\" (UID: \"69b88376-a0bc-4864-bdbb-a445ca0c90b4\") " pod="service-telemetry/stf-smoketest-smoke1-m46pn" Oct 08 00:30:15 crc kubenswrapper[4967]: I1008 00:30:15.318098 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/69b88376-a0bc-4864-bdbb-a445ca0c90b4-ceilometer-entrypoint-script\") pod \"stf-smoketest-smoke1-m46pn\" (UID: \"69b88376-a0bc-4864-bdbb-a445ca0c90b4\") " pod="service-telemetry/stf-smoketest-smoke1-m46pn" Oct 08 00:30:15 crc kubenswrapper[4967]: I1008 00:30:15.346620 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gq8mp\" (UniqueName: \"kubernetes.io/projected/69b88376-a0bc-4864-bdbb-a445ca0c90b4-kube-api-access-gq8mp\") pod \"stf-smoketest-smoke1-m46pn\" (UID: \"69b88376-a0bc-4864-bdbb-a445ca0c90b4\") " pod="service-telemetry/stf-smoketest-smoke1-m46pn" Oct 08 00:30:15 crc kubenswrapper[4967]: I1008 00:30:15.447939 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/stf-smoketest-smoke1-m46pn" Oct 08 00:30:15 crc kubenswrapper[4967]: I1008 00:30:15.579399 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/curl"] Oct 08 00:30:15 crc kubenswrapper[4967]: I1008 00:30:15.603519 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/curl"] Oct 08 00:30:15 crc kubenswrapper[4967]: I1008 00:30:15.603624 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/curl" Oct 08 00:30:15 crc kubenswrapper[4967]: I1008 00:30:15.717754 4967 generic.go:334] "Generic (PLEG): container finished" podID="dd3320fb-5873-4815-ab64-c7fe697c0532" containerID="b25422ee76ce98a45191d83b9fedcea09379d97565986640b5117b3e58cc1aff" exitCode=0 Oct 08 00:30:15 crc kubenswrapper[4967]: I1008 00:30:15.718539 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jk962" event={"ID":"dd3320fb-5873-4815-ab64-c7fe697c0532","Type":"ContainerDied","Data":"b25422ee76ce98a45191d83b9fedcea09379d97565986640b5117b3e58cc1aff"} Oct 08 00:30:15 crc kubenswrapper[4967]: I1008 00:30:15.728483 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-spqfk\" (UniqueName: \"kubernetes.io/projected/8c260cd9-4778-4fad-a61a-c2d46a1c877c-kube-api-access-spqfk\") pod \"curl\" (UID: \"8c260cd9-4778-4fad-a61a-c2d46a1c877c\") " pod="service-telemetry/curl" Oct 08 00:30:15 crc kubenswrapper[4967]: I1008 00:30:15.808826 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/stf-smoketest-smoke1-m46pn"] Oct 08 00:30:15 crc kubenswrapper[4967]: I1008 00:30:15.830961 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-spqfk\" (UniqueName: \"kubernetes.io/projected/8c260cd9-4778-4fad-a61a-c2d46a1c877c-kube-api-access-spqfk\") pod \"curl\" (UID: \"8c260cd9-4778-4fad-a61a-c2d46a1c877c\") " pod="service-telemetry/curl" Oct 08 00:30:15 crc kubenswrapper[4967]: I1008 00:30:15.863395 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-spqfk\" (UniqueName: \"kubernetes.io/projected/8c260cd9-4778-4fad-a61a-c2d46a1c877c-kube-api-access-spqfk\") pod \"curl\" (UID: \"8c260cd9-4778-4fad-a61a-c2d46a1c877c\") " pod="service-telemetry/curl" Oct 08 00:30:15 crc kubenswrapper[4967]: I1008 00:30:15.893916 4967 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-txtvn" podUID="6b156a9a-950b-40da-b51e-e9774ce7b027" containerName="registry-server" probeResult="failure" output=< Oct 08 00:30:15 crc kubenswrapper[4967]: timeout: failed to connect service ":50051" within 1s Oct 08 00:30:15 crc kubenswrapper[4967]: > Oct 08 00:30:15 crc kubenswrapper[4967]: I1008 00:30:15.964636 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/curl" Oct 08 00:30:16 crc kubenswrapper[4967]: I1008 00:30:16.129916 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5f8e39c1-cde6-4742-9d2a-cfbb47b5a14b" path="/var/lib/kubelet/pods/5f8e39c1-cde6-4742-9d2a-cfbb47b5a14b/volumes" Oct 08 00:30:16 crc kubenswrapper[4967]: I1008 00:30:16.418824 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/curl"] Oct 08 00:30:16 crc kubenswrapper[4967]: I1008 00:30:16.725459 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/stf-smoketest-smoke1-m46pn" event={"ID":"69b88376-a0bc-4864-bdbb-a445ca0c90b4","Type":"ContainerStarted","Data":"52ab402b474cc2c62c606a702f6f591708724a99c6a447a96030ed5da53c7cec"} Oct 08 00:30:16 crc kubenswrapper[4967]: I1008 00:30:16.727557 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/curl" event={"ID":"8c260cd9-4778-4fad-a61a-c2d46a1c877c","Type":"ContainerStarted","Data":"819b6e2a02a51f4a567ae8ff702c983c5f796779fa17fca71065c772e37f8f0f"} Oct 08 00:30:18 crc kubenswrapper[4967]: I1008 00:30:18.742947 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jk962" event={"ID":"dd3320fb-5873-4815-ab64-c7fe697c0532","Type":"ContainerStarted","Data":"258ba7d3a857e5aa6c65e2121d2248309a08535d93bbe03cb911a58fe65a827c"} Oct 08 00:30:19 crc kubenswrapper[4967]: I1008 00:30:19.758078 4967 generic.go:334] "Generic (PLEG): container finished" podID="dd3320fb-5873-4815-ab64-c7fe697c0532" containerID="258ba7d3a857e5aa6c65e2121d2248309a08535d93bbe03cb911a58fe65a827c" exitCode=0 Oct 08 00:30:19 crc kubenswrapper[4967]: I1008 00:30:19.758120 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jk962" event={"ID":"dd3320fb-5873-4815-ab64-c7fe697c0532","Type":"ContainerDied","Data":"258ba7d3a857e5aa6c65e2121d2248309a08535d93bbe03cb911a58fe65a827c"} Oct 08 00:30:20 crc kubenswrapper[4967]: I1008 00:30:20.765211 4967 generic.go:334] "Generic (PLEG): container finished" podID="8c260cd9-4778-4fad-a61a-c2d46a1c877c" containerID="bb7293d08c61850f3f7b1984c5b9ad2f00e5954a054d46a063a4093bae0987c0" exitCode=0 Oct 08 00:30:20 crc kubenswrapper[4967]: I1008 00:30:20.765826 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/curl" event={"ID":"8c260cd9-4778-4fad-a61a-c2d46a1c877c","Type":"ContainerDied","Data":"bb7293d08c61850f3f7b1984c5b9ad2f00e5954a054d46a063a4093bae0987c0"} Oct 08 00:30:20 crc kubenswrapper[4967]: I1008 00:30:20.768602 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jk962" event={"ID":"dd3320fb-5873-4815-ab64-c7fe697c0532","Type":"ContainerStarted","Data":"c9bf0cfb03e3ac5c0e1add64a4d441a39f89eb88e73ae6d4f2a34202e3864175"} Oct 08 00:30:20 crc kubenswrapper[4967]: I1008 00:30:20.797203 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-jk962" podStartSLOduration=3.957615685 podStartE2EDuration="9.79718793s" podCreationTimestamp="2025-10-08 00:30:11 +0000 UTC" firstStartedPulling="2025-10-08 00:30:14.665827018 +0000 UTC m=+1413.316361667" lastFinishedPulling="2025-10-08 00:30:20.505399263 +0000 UTC m=+1419.155933912" observedRunningTime="2025-10-08 00:30:20.793125318 +0000 UTC m=+1419.443659977" watchObservedRunningTime="2025-10-08 00:30:20.79718793 +0000 UTC m=+1419.447722579" Oct 08 00:30:21 crc kubenswrapper[4967]: I1008 00:30:21.878465 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-jk962" Oct 08 00:30:21 crc kubenswrapper[4967]: I1008 00:30:21.878920 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-jk962" Oct 08 00:30:22 crc kubenswrapper[4967]: I1008 00:30:22.966283 4967 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-jk962" podUID="dd3320fb-5873-4815-ab64-c7fe697c0532" containerName="registry-server" probeResult="failure" output=< Oct 08 00:30:22 crc kubenswrapper[4967]: timeout: failed to connect service ":50051" within 1s Oct 08 00:30:22 crc kubenswrapper[4967]: > Oct 08 00:30:23 crc kubenswrapper[4967]: I1008 00:30:23.002290 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/curl" Oct 08 00:30:23 crc kubenswrapper[4967]: I1008 00:30:23.167618 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-spqfk\" (UniqueName: \"kubernetes.io/projected/8c260cd9-4778-4fad-a61a-c2d46a1c877c-kube-api-access-spqfk\") pod \"8c260cd9-4778-4fad-a61a-c2d46a1c877c\" (UID: \"8c260cd9-4778-4fad-a61a-c2d46a1c877c\") " Oct 08 00:30:23 crc kubenswrapper[4967]: I1008 00:30:23.172384 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c260cd9-4778-4fad-a61a-c2d46a1c877c-kube-api-access-spqfk" (OuterVolumeSpecName: "kube-api-access-spqfk") pod "8c260cd9-4778-4fad-a61a-c2d46a1c877c" (UID: "8c260cd9-4778-4fad-a61a-c2d46a1c877c"). InnerVolumeSpecName "kube-api-access-spqfk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:30:23 crc kubenswrapper[4967]: I1008 00:30:23.232529 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_curl_8c260cd9-4778-4fad-a61a-c2d46a1c877c/curl/0.log" Oct 08 00:30:23 crc kubenswrapper[4967]: I1008 00:30:23.269434 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-spqfk\" (UniqueName: \"kubernetes.io/projected/8c260cd9-4778-4fad-a61a-c2d46a1c877c-kube-api-access-spqfk\") on node \"crc\" DevicePath \"\"" Oct 08 00:30:23 crc kubenswrapper[4967]: I1008 00:30:23.589544 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-snmp-webhook-6856cfb745-6jgkj_f3eb705e-460b-45f4-be46-8d811bda63c3/prometheus-webhook-snmp/0.log" Oct 08 00:30:23 crc kubenswrapper[4967]: I1008 00:30:23.798005 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/curl" Oct 08 00:30:23 crc kubenswrapper[4967]: I1008 00:30:23.798383 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/curl" event={"ID":"8c260cd9-4778-4fad-a61a-c2d46a1c877c","Type":"ContainerDied","Data":"819b6e2a02a51f4a567ae8ff702c983c5f796779fa17fca71065c772e37f8f0f"} Oct 08 00:30:23 crc kubenswrapper[4967]: I1008 00:30:23.798404 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="819b6e2a02a51f4a567ae8ff702c983c5f796779fa17fca71065c772e37f8f0f" Oct 08 00:30:24 crc kubenswrapper[4967]: I1008 00:30:24.979024 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-txtvn" Oct 08 00:30:25 crc kubenswrapper[4967]: I1008 00:30:25.022492 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-txtvn" Oct 08 00:30:25 crc kubenswrapper[4967]: I1008 00:30:25.207649 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-txtvn"] Oct 08 00:30:26 crc kubenswrapper[4967]: I1008 00:30:26.820381 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-txtvn" podUID="6b156a9a-950b-40da-b51e-e9774ce7b027" containerName="registry-server" containerID="cri-o://fe2fdead628193cfbf79bf235241329ed1ed7b11cf34ad489a04a0ebee862660" gracePeriod=2 Oct 08 00:30:27 crc kubenswrapper[4967]: I1008 00:30:27.829314 4967 generic.go:334] "Generic (PLEG): container finished" podID="6b156a9a-950b-40da-b51e-e9774ce7b027" containerID="fe2fdead628193cfbf79bf235241329ed1ed7b11cf34ad489a04a0ebee862660" exitCode=0 Oct 08 00:30:27 crc kubenswrapper[4967]: I1008 00:30:27.829362 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-txtvn" event={"ID":"6b156a9a-950b-40da-b51e-e9774ce7b027","Type":"ContainerDied","Data":"fe2fdead628193cfbf79bf235241329ed1ed7b11cf34ad489a04a0ebee862660"} Oct 08 00:30:31 crc kubenswrapper[4967]: I1008 00:30:31.756736 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-txtvn" Oct 08 00:30:31 crc kubenswrapper[4967]: I1008 00:30:31.866167 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-txtvn" event={"ID":"6b156a9a-950b-40da-b51e-e9774ce7b027","Type":"ContainerDied","Data":"8a8da8f4c2dca8a8072b541f1eae6d865af9e930337fe713b23099c5fd630671"} Oct 08 00:30:31 crc kubenswrapper[4967]: I1008 00:30:31.866539 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-txtvn" Oct 08 00:30:31 crc kubenswrapper[4967]: I1008 00:30:31.866662 4967 scope.go:117] "RemoveContainer" containerID="fe2fdead628193cfbf79bf235241329ed1ed7b11cf34ad489a04a0ebee862660" Oct 08 00:30:31 crc kubenswrapper[4967]: I1008 00:30:31.872219 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/stf-smoketest-smoke1-m46pn" event={"ID":"69b88376-a0bc-4864-bdbb-a445ca0c90b4","Type":"ContainerStarted","Data":"5adf8df03ecb4a32425819c620ad490cedd0c977d0afc0f1b6e520a569bbdf39"} Oct 08 00:30:31 crc kubenswrapper[4967]: I1008 00:30:31.886155 4967 scope.go:117] "RemoveContainer" containerID="589d45040ff6bcdc265de3e03ee5be21ef82d1cd3ce65a30de12e71ca59ba3f7" Oct 08 00:30:31 crc kubenswrapper[4967]: I1008 00:30:31.905118 4967 scope.go:117] "RemoveContainer" containerID="6658ff11c323ece9282af45a9d61f28182be1130e3e13a92e006b90e43287621" Oct 08 00:30:31 crc kubenswrapper[4967]: I1008 00:30:31.906769 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b156a9a-950b-40da-b51e-e9774ce7b027-catalog-content\") pod \"6b156a9a-950b-40da-b51e-e9774ce7b027\" (UID: \"6b156a9a-950b-40da-b51e-e9774ce7b027\") " Oct 08 00:30:31 crc kubenswrapper[4967]: I1008 00:30:31.906868 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hzrbl\" (UniqueName: \"kubernetes.io/projected/6b156a9a-950b-40da-b51e-e9774ce7b027-kube-api-access-hzrbl\") pod \"6b156a9a-950b-40da-b51e-e9774ce7b027\" (UID: \"6b156a9a-950b-40da-b51e-e9774ce7b027\") " Oct 08 00:30:31 crc kubenswrapper[4967]: I1008 00:30:31.907036 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b156a9a-950b-40da-b51e-e9774ce7b027-utilities\") pod \"6b156a9a-950b-40da-b51e-e9774ce7b027\" (UID: \"6b156a9a-950b-40da-b51e-e9774ce7b027\") " Oct 08 00:30:31 crc kubenswrapper[4967]: I1008 00:30:31.908363 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b156a9a-950b-40da-b51e-e9774ce7b027-utilities" (OuterVolumeSpecName: "utilities") pod "6b156a9a-950b-40da-b51e-e9774ce7b027" (UID: "6b156a9a-950b-40da-b51e-e9774ce7b027"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:30:31 crc kubenswrapper[4967]: I1008 00:30:31.912699 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b156a9a-950b-40da-b51e-e9774ce7b027-kube-api-access-hzrbl" (OuterVolumeSpecName: "kube-api-access-hzrbl") pod "6b156a9a-950b-40da-b51e-e9774ce7b027" (UID: "6b156a9a-950b-40da-b51e-e9774ce7b027"). InnerVolumeSpecName "kube-api-access-hzrbl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:30:31 crc kubenswrapper[4967]: I1008 00:30:31.934925 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-jk962" Oct 08 00:30:31 crc kubenswrapper[4967]: I1008 00:30:31.952655 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b156a9a-950b-40da-b51e-e9774ce7b027-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6b156a9a-950b-40da-b51e-e9774ce7b027" (UID: "6b156a9a-950b-40da-b51e-e9774ce7b027"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:30:32 crc kubenswrapper[4967]: I1008 00:30:32.000186 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-jk962" Oct 08 00:30:32 crc kubenswrapper[4967]: I1008 00:30:32.009748 4967 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b156a9a-950b-40da-b51e-e9774ce7b027-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 00:30:32 crc kubenswrapper[4967]: I1008 00:30:32.009785 4967 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b156a9a-950b-40da-b51e-e9774ce7b027-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 00:30:32 crc kubenswrapper[4967]: I1008 00:30:32.009799 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hzrbl\" (UniqueName: \"kubernetes.io/projected/6b156a9a-950b-40da-b51e-e9774ce7b027-kube-api-access-hzrbl\") on node \"crc\" DevicePath \"\"" Oct 08 00:30:32 crc kubenswrapper[4967]: I1008 00:30:32.183608 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-txtvn"] Oct 08 00:30:32 crc kubenswrapper[4967]: I1008 00:30:32.188419 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-txtvn"] Oct 08 00:30:33 crc kubenswrapper[4967]: I1008 00:30:33.586491 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jk962"] Oct 08 00:30:33 crc kubenswrapper[4967]: I1008 00:30:33.901811 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-jk962" podUID="dd3320fb-5873-4815-ab64-c7fe697c0532" containerName="registry-server" containerID="cri-o://c9bf0cfb03e3ac5c0e1add64a4d441a39f89eb88e73ae6d4f2a34202e3864175" gracePeriod=2 Oct 08 00:30:34 crc kubenswrapper[4967]: I1008 00:30:34.132017 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b156a9a-950b-40da-b51e-e9774ce7b027" path="/var/lib/kubelet/pods/6b156a9a-950b-40da-b51e-e9774ce7b027/volumes" Oct 08 00:30:34 crc kubenswrapper[4967]: I1008 00:30:34.914567 4967 generic.go:334] "Generic (PLEG): container finished" podID="dd3320fb-5873-4815-ab64-c7fe697c0532" containerID="c9bf0cfb03e3ac5c0e1add64a4d441a39f89eb88e73ae6d4f2a34202e3864175" exitCode=0 Oct 08 00:30:34 crc kubenswrapper[4967]: I1008 00:30:34.915568 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jk962" event={"ID":"dd3320fb-5873-4815-ab64-c7fe697c0532","Type":"ContainerDied","Data":"c9bf0cfb03e3ac5c0e1add64a4d441a39f89eb88e73ae6d4f2a34202e3864175"} Oct 08 00:30:36 crc kubenswrapper[4967]: I1008 00:30:36.920824 4967 patch_prober.go:28] interesting pod/machine-config-daemon-xvdtp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 00:30:36 crc kubenswrapper[4967]: I1008 00:30:36.921153 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" podUID="9461162a-c27d-465e-9ca7-2d3936877cd8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 00:30:37 crc kubenswrapper[4967]: I1008 00:30:37.476084 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jk962" Oct 08 00:30:37 crc kubenswrapper[4967]: I1008 00:30:37.601367 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dqqbd\" (UniqueName: \"kubernetes.io/projected/dd3320fb-5873-4815-ab64-c7fe697c0532-kube-api-access-dqqbd\") pod \"dd3320fb-5873-4815-ab64-c7fe697c0532\" (UID: \"dd3320fb-5873-4815-ab64-c7fe697c0532\") " Oct 08 00:30:37 crc kubenswrapper[4967]: I1008 00:30:37.601443 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd3320fb-5873-4815-ab64-c7fe697c0532-utilities\") pod \"dd3320fb-5873-4815-ab64-c7fe697c0532\" (UID: \"dd3320fb-5873-4815-ab64-c7fe697c0532\") " Oct 08 00:30:37 crc kubenswrapper[4967]: I1008 00:30:37.601480 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd3320fb-5873-4815-ab64-c7fe697c0532-catalog-content\") pod \"dd3320fb-5873-4815-ab64-c7fe697c0532\" (UID: \"dd3320fb-5873-4815-ab64-c7fe697c0532\") " Oct 08 00:30:37 crc kubenswrapper[4967]: I1008 00:30:37.603137 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dd3320fb-5873-4815-ab64-c7fe697c0532-utilities" (OuterVolumeSpecName: "utilities") pod "dd3320fb-5873-4815-ab64-c7fe697c0532" (UID: "dd3320fb-5873-4815-ab64-c7fe697c0532"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:30:37 crc kubenswrapper[4967]: I1008 00:30:37.609825 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd3320fb-5873-4815-ab64-c7fe697c0532-kube-api-access-dqqbd" (OuterVolumeSpecName: "kube-api-access-dqqbd") pod "dd3320fb-5873-4815-ab64-c7fe697c0532" (UID: "dd3320fb-5873-4815-ab64-c7fe697c0532"). InnerVolumeSpecName "kube-api-access-dqqbd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:30:37 crc kubenswrapper[4967]: I1008 00:30:37.664315 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dd3320fb-5873-4815-ab64-c7fe697c0532-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dd3320fb-5873-4815-ab64-c7fe697c0532" (UID: "dd3320fb-5873-4815-ab64-c7fe697c0532"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:30:37 crc kubenswrapper[4967]: I1008 00:30:37.702751 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dqqbd\" (UniqueName: \"kubernetes.io/projected/dd3320fb-5873-4815-ab64-c7fe697c0532-kube-api-access-dqqbd\") on node \"crc\" DevicePath \"\"" Oct 08 00:30:37 crc kubenswrapper[4967]: I1008 00:30:37.702780 4967 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd3320fb-5873-4815-ab64-c7fe697c0532-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 00:30:37 crc kubenswrapper[4967]: I1008 00:30:37.702789 4967 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd3320fb-5873-4815-ab64-c7fe697c0532-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 00:30:37 crc kubenswrapper[4967]: I1008 00:30:37.943701 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jk962" event={"ID":"dd3320fb-5873-4815-ab64-c7fe697c0532","Type":"ContainerDied","Data":"c872194e3b5ad502f28189aafe5272036fad0ebb02478f09c37d0c4921b92514"} Oct 08 00:30:37 crc kubenswrapper[4967]: I1008 00:30:37.943797 4967 scope.go:117] "RemoveContainer" containerID="c9bf0cfb03e3ac5c0e1add64a4d441a39f89eb88e73ae6d4f2a34202e3864175" Oct 08 00:30:37 crc kubenswrapper[4967]: I1008 00:30:37.943749 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jk962" Oct 08 00:30:37 crc kubenswrapper[4967]: I1008 00:30:37.946587 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/stf-smoketest-smoke1-m46pn" event={"ID":"69b88376-a0bc-4864-bdbb-a445ca0c90b4","Type":"ContainerStarted","Data":"4384fb871cbae65d9aa2e87a0bb1c2b87db4d75449acb7df0858b111a1e001f4"} Oct 08 00:30:37 crc kubenswrapper[4967]: I1008 00:30:37.973878 4967 scope.go:117] "RemoveContainer" containerID="258ba7d3a857e5aa6c65e2121d2248309a08535d93bbe03cb911a58fe65a827c" Oct 08 00:30:37 crc kubenswrapper[4967]: I1008 00:30:37.978342 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/stf-smoketest-smoke1-m46pn" podStartSLOduration=1.339772342 podStartE2EDuration="22.978320784s" podCreationTimestamp="2025-10-08 00:30:15 +0000 UTC" firstStartedPulling="2025-10-08 00:30:15.831706506 +0000 UTC m=+1414.482241155" lastFinishedPulling="2025-10-08 00:30:37.470254938 +0000 UTC m=+1436.120789597" observedRunningTime="2025-10-08 00:30:37.977858631 +0000 UTC m=+1436.628393370" watchObservedRunningTime="2025-10-08 00:30:37.978320784 +0000 UTC m=+1436.628855433" Oct 08 00:30:38 crc kubenswrapper[4967]: I1008 00:30:38.004889 4967 scope.go:117] "RemoveContainer" containerID="b25422ee76ce98a45191d83b9fedcea09379d97565986640b5117b3e58cc1aff" Oct 08 00:30:38 crc kubenswrapper[4967]: I1008 00:30:38.013878 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jk962"] Oct 08 00:30:38 crc kubenswrapper[4967]: I1008 00:30:38.036182 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-jk962"] Oct 08 00:30:38 crc kubenswrapper[4967]: I1008 00:30:38.131207 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd3320fb-5873-4815-ab64-c7fe697c0532" path="/var/lib/kubelet/pods/dd3320fb-5873-4815-ab64-c7fe697c0532/volumes" Oct 08 00:30:53 crc kubenswrapper[4967]: I1008 00:30:53.774925 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-snmp-webhook-6856cfb745-6jgkj_f3eb705e-460b-45f4-be46-8d811bda63c3/prometheus-webhook-snmp/0.log" Oct 08 00:31:06 crc kubenswrapper[4967]: I1008 00:31:06.239071 4967 generic.go:334] "Generic (PLEG): container finished" podID="69b88376-a0bc-4864-bdbb-a445ca0c90b4" containerID="5adf8df03ecb4a32425819c620ad490cedd0c977d0afc0f1b6e520a569bbdf39" exitCode=1 Oct 08 00:31:06 crc kubenswrapper[4967]: I1008 00:31:06.239261 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/stf-smoketest-smoke1-m46pn" event={"ID":"69b88376-a0bc-4864-bdbb-a445ca0c90b4","Type":"ContainerDied","Data":"5adf8df03ecb4a32425819c620ad490cedd0c977d0afc0f1b6e520a569bbdf39"} Oct 08 00:31:06 crc kubenswrapper[4967]: I1008 00:31:06.240644 4967 scope.go:117] "RemoveContainer" containerID="5adf8df03ecb4a32425819c620ad490cedd0c977d0afc0f1b6e520a569bbdf39" Oct 08 00:31:06 crc kubenswrapper[4967]: I1008 00:31:06.920479 4967 patch_prober.go:28] interesting pod/machine-config-daemon-xvdtp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 00:31:06 crc kubenswrapper[4967]: I1008 00:31:06.920551 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" podUID="9461162a-c27d-465e-9ca7-2d3936877cd8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 00:31:06 crc kubenswrapper[4967]: I1008 00:31:06.920598 4967 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" Oct 08 00:31:06 crc kubenswrapper[4967]: I1008 00:31:06.921183 4967 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5e1e0751dcae2b0e4bad8c7f2220f3bca21a69c9cccc3dd960fa9770bb1adf17"} pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 00:31:06 crc kubenswrapper[4967]: I1008 00:31:06.921239 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" podUID="9461162a-c27d-465e-9ca7-2d3936877cd8" containerName="machine-config-daemon" containerID="cri-o://5e1e0751dcae2b0e4bad8c7f2220f3bca21a69c9cccc3dd960fa9770bb1adf17" gracePeriod=600 Oct 08 00:31:07 crc kubenswrapper[4967]: I1008 00:31:07.254911 4967 generic.go:334] "Generic (PLEG): container finished" podID="9461162a-c27d-465e-9ca7-2d3936877cd8" containerID="5e1e0751dcae2b0e4bad8c7f2220f3bca21a69c9cccc3dd960fa9770bb1adf17" exitCode=0 Oct 08 00:31:07 crc kubenswrapper[4967]: I1008 00:31:07.255005 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" event={"ID":"9461162a-c27d-465e-9ca7-2d3936877cd8","Type":"ContainerDied","Data":"5e1e0751dcae2b0e4bad8c7f2220f3bca21a69c9cccc3dd960fa9770bb1adf17"} Oct 08 00:31:07 crc kubenswrapper[4967]: I1008 00:31:07.255589 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" event={"ID":"9461162a-c27d-465e-9ca7-2d3936877cd8","Type":"ContainerStarted","Data":"dd4b0abe7436dc2e0c64a27b31f4a959aef1a772ba57812e2a8fbddc81d33d75"} Oct 08 00:31:07 crc kubenswrapper[4967]: I1008 00:31:07.255616 4967 scope.go:117] "RemoveContainer" containerID="a9bcd32e3405379dbe88886537687c4d0982cbf18aecd6bc1e1ae62026f21458" Oct 08 00:31:09 crc kubenswrapper[4967]: I1008 00:31:09.287069 4967 generic.go:334] "Generic (PLEG): container finished" podID="69b88376-a0bc-4864-bdbb-a445ca0c90b4" containerID="4384fb871cbae65d9aa2e87a0bb1c2b87db4d75449acb7df0858b111a1e001f4" exitCode=0 Oct 08 00:31:09 crc kubenswrapper[4967]: I1008 00:31:09.287385 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/stf-smoketest-smoke1-m46pn" event={"ID":"69b88376-a0bc-4864-bdbb-a445ca0c90b4","Type":"ContainerDied","Data":"4384fb871cbae65d9aa2e87a0bb1c2b87db4d75449acb7df0858b111a1e001f4"} Oct 08 00:31:10 crc kubenswrapper[4967]: I1008 00:31:10.668437 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/stf-smoketest-smoke1-m46pn" Oct 08 00:31:10 crc kubenswrapper[4967]: I1008 00:31:10.746026 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"healthcheck-log\" (UniqueName: \"kubernetes.io/configmap/69b88376-a0bc-4864-bdbb-a445ca0c90b4-healthcheck-log\") pod \"69b88376-a0bc-4864-bdbb-a445ca0c90b4\" (UID: \"69b88376-a0bc-4864-bdbb-a445ca0c90b4\") " Oct 08 00:31:10 crc kubenswrapper[4967]: I1008 00:31:10.746114 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"collectd-config\" (UniqueName: \"kubernetes.io/configmap/69b88376-a0bc-4864-bdbb-a445ca0c90b4-collectd-config\") pod \"69b88376-a0bc-4864-bdbb-a445ca0c90b4\" (UID: \"69b88376-a0bc-4864-bdbb-a445ca0c90b4\") " Oct 08 00:31:10 crc kubenswrapper[4967]: I1008 00:31:10.746139 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gq8mp\" (UniqueName: \"kubernetes.io/projected/69b88376-a0bc-4864-bdbb-a445ca0c90b4-kube-api-access-gq8mp\") pod \"69b88376-a0bc-4864-bdbb-a445ca0c90b4\" (UID: \"69b88376-a0bc-4864-bdbb-a445ca0c90b4\") " Oct 08 00:31:10 crc kubenswrapper[4967]: I1008 00:31:10.746165 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/69b88376-a0bc-4864-bdbb-a445ca0c90b4-ceilometer-entrypoint-script\") pod \"69b88376-a0bc-4864-bdbb-a445ca0c90b4\" (UID: \"69b88376-a0bc-4864-bdbb-a445ca0c90b4\") " Oct 08 00:31:10 crc kubenswrapper[4967]: I1008 00:31:10.746204 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sensubility-config\" (UniqueName: \"kubernetes.io/configmap/69b88376-a0bc-4864-bdbb-a445ca0c90b4-sensubility-config\") pod \"69b88376-a0bc-4864-bdbb-a445ca0c90b4\" (UID: \"69b88376-a0bc-4864-bdbb-a445ca0c90b4\") " Oct 08 00:31:10 crc kubenswrapper[4967]: I1008 00:31:10.746228 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"collectd-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/69b88376-a0bc-4864-bdbb-a445ca0c90b4-collectd-entrypoint-script\") pod \"69b88376-a0bc-4864-bdbb-a445ca0c90b4\" (UID: \"69b88376-a0bc-4864-bdbb-a445ca0c90b4\") " Oct 08 00:31:10 crc kubenswrapper[4967]: I1008 00:31:10.746275 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-publisher\" (UniqueName: \"kubernetes.io/configmap/69b88376-a0bc-4864-bdbb-a445ca0c90b4-ceilometer-publisher\") pod \"69b88376-a0bc-4864-bdbb-a445ca0c90b4\" (UID: \"69b88376-a0bc-4864-bdbb-a445ca0c90b4\") " Oct 08 00:31:10 crc kubenswrapper[4967]: I1008 00:31:10.753866 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69b88376-a0bc-4864-bdbb-a445ca0c90b4-kube-api-access-gq8mp" (OuterVolumeSpecName: "kube-api-access-gq8mp") pod "69b88376-a0bc-4864-bdbb-a445ca0c90b4" (UID: "69b88376-a0bc-4864-bdbb-a445ca0c90b4"). InnerVolumeSpecName "kube-api-access-gq8mp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:31:10 crc kubenswrapper[4967]: I1008 00:31:10.770601 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69b88376-a0bc-4864-bdbb-a445ca0c90b4-sensubility-config" (OuterVolumeSpecName: "sensubility-config") pod "69b88376-a0bc-4864-bdbb-a445ca0c90b4" (UID: "69b88376-a0bc-4864-bdbb-a445ca0c90b4"). InnerVolumeSpecName "sensubility-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:31:10 crc kubenswrapper[4967]: I1008 00:31:10.775839 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69b88376-a0bc-4864-bdbb-a445ca0c90b4-ceilometer-entrypoint-script" (OuterVolumeSpecName: "ceilometer-entrypoint-script") pod "69b88376-a0bc-4864-bdbb-a445ca0c90b4" (UID: "69b88376-a0bc-4864-bdbb-a445ca0c90b4"). InnerVolumeSpecName "ceilometer-entrypoint-script". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:31:10 crc kubenswrapper[4967]: I1008 00:31:10.777734 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69b88376-a0bc-4864-bdbb-a445ca0c90b4-ceilometer-publisher" (OuterVolumeSpecName: "ceilometer-publisher") pod "69b88376-a0bc-4864-bdbb-a445ca0c90b4" (UID: "69b88376-a0bc-4864-bdbb-a445ca0c90b4"). InnerVolumeSpecName "ceilometer-publisher". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:31:10 crc kubenswrapper[4967]: I1008 00:31:10.777890 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69b88376-a0bc-4864-bdbb-a445ca0c90b4-healthcheck-log" (OuterVolumeSpecName: "healthcheck-log") pod "69b88376-a0bc-4864-bdbb-a445ca0c90b4" (UID: "69b88376-a0bc-4864-bdbb-a445ca0c90b4"). InnerVolumeSpecName "healthcheck-log". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:31:10 crc kubenswrapper[4967]: I1008 00:31:10.785791 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69b88376-a0bc-4864-bdbb-a445ca0c90b4-collectd-config" (OuterVolumeSpecName: "collectd-config") pod "69b88376-a0bc-4864-bdbb-a445ca0c90b4" (UID: "69b88376-a0bc-4864-bdbb-a445ca0c90b4"). InnerVolumeSpecName "collectd-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:31:10 crc kubenswrapper[4967]: I1008 00:31:10.789225 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69b88376-a0bc-4864-bdbb-a445ca0c90b4-collectd-entrypoint-script" (OuterVolumeSpecName: "collectd-entrypoint-script") pod "69b88376-a0bc-4864-bdbb-a445ca0c90b4" (UID: "69b88376-a0bc-4864-bdbb-a445ca0c90b4"). InnerVolumeSpecName "collectd-entrypoint-script". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:31:10 crc kubenswrapper[4967]: I1008 00:31:10.848608 4967 reconciler_common.go:293] "Volume detached for volume \"sensubility-config\" (UniqueName: \"kubernetes.io/configmap/69b88376-a0bc-4864-bdbb-a445ca0c90b4-sensubility-config\") on node \"crc\" DevicePath \"\"" Oct 08 00:31:10 crc kubenswrapper[4967]: I1008 00:31:10.848654 4967 reconciler_common.go:293] "Volume detached for volume \"collectd-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/69b88376-a0bc-4864-bdbb-a445ca0c90b4-collectd-entrypoint-script\") on node \"crc\" DevicePath \"\"" Oct 08 00:31:10 crc kubenswrapper[4967]: I1008 00:31:10.848682 4967 reconciler_common.go:293] "Volume detached for volume \"ceilometer-publisher\" (UniqueName: \"kubernetes.io/configmap/69b88376-a0bc-4864-bdbb-a445ca0c90b4-ceilometer-publisher\") on node \"crc\" DevicePath \"\"" Oct 08 00:31:10 crc kubenswrapper[4967]: I1008 00:31:10.848693 4967 reconciler_common.go:293] "Volume detached for volume \"healthcheck-log\" (UniqueName: \"kubernetes.io/configmap/69b88376-a0bc-4864-bdbb-a445ca0c90b4-healthcheck-log\") on node \"crc\" DevicePath \"\"" Oct 08 00:31:10 crc kubenswrapper[4967]: I1008 00:31:10.848701 4967 reconciler_common.go:293] "Volume detached for volume \"collectd-config\" (UniqueName: \"kubernetes.io/configmap/69b88376-a0bc-4864-bdbb-a445ca0c90b4-collectd-config\") on node \"crc\" DevicePath \"\"" Oct 08 00:31:10 crc kubenswrapper[4967]: I1008 00:31:10.848712 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gq8mp\" (UniqueName: \"kubernetes.io/projected/69b88376-a0bc-4864-bdbb-a445ca0c90b4-kube-api-access-gq8mp\") on node \"crc\" DevicePath \"\"" Oct 08 00:31:10 crc kubenswrapper[4967]: I1008 00:31:10.848721 4967 reconciler_common.go:293] "Volume detached for volume \"ceilometer-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/69b88376-a0bc-4864-bdbb-a445ca0c90b4-ceilometer-entrypoint-script\") on node \"crc\" DevicePath \"\"" Oct 08 00:31:11 crc kubenswrapper[4967]: I1008 00:31:11.306116 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/stf-smoketest-smoke1-m46pn" event={"ID":"69b88376-a0bc-4864-bdbb-a445ca0c90b4","Type":"ContainerDied","Data":"52ab402b474cc2c62c606a702f6f591708724a99c6a447a96030ed5da53c7cec"} Oct 08 00:31:11 crc kubenswrapper[4967]: I1008 00:31:11.306171 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="52ab402b474cc2c62c606a702f6f591708724a99c6a447a96030ed5da53c7cec" Oct 08 00:31:11 crc kubenswrapper[4967]: I1008 00:31:11.306481 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/stf-smoketest-smoke1-m46pn" Oct 08 00:31:19 crc kubenswrapper[4967]: I1008 00:31:19.046009 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/stf-smoketest-smoke1-6s297"] Oct 08 00:31:19 crc kubenswrapper[4967]: E1008 00:31:19.047120 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69b88376-a0bc-4864-bdbb-a445ca0c90b4" containerName="smoketest-ceilometer" Oct 08 00:31:19 crc kubenswrapper[4967]: I1008 00:31:19.047150 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="69b88376-a0bc-4864-bdbb-a445ca0c90b4" containerName="smoketest-ceilometer" Oct 08 00:31:19 crc kubenswrapper[4967]: E1008 00:31:19.047188 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd3320fb-5873-4815-ab64-c7fe697c0532" containerName="extract-utilities" Oct 08 00:31:19 crc kubenswrapper[4967]: I1008 00:31:19.047206 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd3320fb-5873-4815-ab64-c7fe697c0532" containerName="extract-utilities" Oct 08 00:31:19 crc kubenswrapper[4967]: E1008 00:31:19.047232 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69b88376-a0bc-4864-bdbb-a445ca0c90b4" containerName="smoketest-collectd" Oct 08 00:31:19 crc kubenswrapper[4967]: I1008 00:31:19.047248 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="69b88376-a0bc-4864-bdbb-a445ca0c90b4" containerName="smoketest-collectd" Oct 08 00:31:19 crc kubenswrapper[4967]: E1008 00:31:19.047267 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd3320fb-5873-4815-ab64-c7fe697c0532" containerName="extract-content" Oct 08 00:31:19 crc kubenswrapper[4967]: I1008 00:31:19.047282 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd3320fb-5873-4815-ab64-c7fe697c0532" containerName="extract-content" Oct 08 00:31:19 crc kubenswrapper[4967]: E1008 00:31:19.047323 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b156a9a-950b-40da-b51e-e9774ce7b027" containerName="registry-server" Oct 08 00:31:19 crc kubenswrapper[4967]: I1008 00:31:19.047340 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b156a9a-950b-40da-b51e-e9774ce7b027" containerName="registry-server" Oct 08 00:31:19 crc kubenswrapper[4967]: E1008 00:31:19.047366 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd3320fb-5873-4815-ab64-c7fe697c0532" containerName="registry-server" Oct 08 00:31:19 crc kubenswrapper[4967]: I1008 00:31:19.047384 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd3320fb-5873-4815-ab64-c7fe697c0532" containerName="registry-server" Oct 08 00:31:19 crc kubenswrapper[4967]: E1008 00:31:19.047402 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b156a9a-950b-40da-b51e-e9774ce7b027" containerName="extract-utilities" Oct 08 00:31:19 crc kubenswrapper[4967]: I1008 00:31:19.047419 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b156a9a-950b-40da-b51e-e9774ce7b027" containerName="extract-utilities" Oct 08 00:31:19 crc kubenswrapper[4967]: E1008 00:31:19.047445 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c260cd9-4778-4fad-a61a-c2d46a1c877c" containerName="curl" Oct 08 00:31:19 crc kubenswrapper[4967]: I1008 00:31:19.047460 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c260cd9-4778-4fad-a61a-c2d46a1c877c" containerName="curl" Oct 08 00:31:19 crc kubenswrapper[4967]: E1008 00:31:19.047484 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b156a9a-950b-40da-b51e-e9774ce7b027" containerName="extract-content" Oct 08 00:31:19 crc kubenswrapper[4967]: I1008 00:31:19.047501 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b156a9a-950b-40da-b51e-e9774ce7b027" containerName="extract-content" Oct 08 00:31:19 crc kubenswrapper[4967]: I1008 00:31:19.047811 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="69b88376-a0bc-4864-bdbb-a445ca0c90b4" containerName="smoketest-ceilometer" Oct 08 00:31:19 crc kubenswrapper[4967]: I1008 00:31:19.047834 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd3320fb-5873-4815-ab64-c7fe697c0532" containerName="registry-server" Oct 08 00:31:19 crc kubenswrapper[4967]: I1008 00:31:19.047863 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c260cd9-4778-4fad-a61a-c2d46a1c877c" containerName="curl" Oct 08 00:31:19 crc kubenswrapper[4967]: I1008 00:31:19.047889 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b156a9a-950b-40da-b51e-e9774ce7b027" containerName="registry-server" Oct 08 00:31:19 crc kubenswrapper[4967]: I1008 00:31:19.047906 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="69b88376-a0bc-4864-bdbb-a445ca0c90b4" containerName="smoketest-collectd" Oct 08 00:31:19 crc kubenswrapper[4967]: I1008 00:31:19.049272 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/stf-smoketest-smoke1-6s297" Oct 08 00:31:19 crc kubenswrapper[4967]: I1008 00:31:19.052704 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"stf-smoketest-ceilometer-entrypoint-script" Oct 08 00:31:19 crc kubenswrapper[4967]: I1008 00:31:19.053224 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"stf-smoketest-sensubility-config" Oct 08 00:31:19 crc kubenswrapper[4967]: I1008 00:31:19.053853 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"stf-smoketest-healthcheck-log" Oct 08 00:31:19 crc kubenswrapper[4967]: I1008 00:31:19.054346 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"stf-smoketest-collectd-entrypoint-script" Oct 08 00:31:19 crc kubenswrapper[4967]: I1008 00:31:19.054589 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"stf-smoketest-collectd-config" Oct 08 00:31:19 crc kubenswrapper[4967]: I1008 00:31:19.055563 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"stf-smoketest-ceilometer-publisher" Oct 08 00:31:19 crc kubenswrapper[4967]: I1008 00:31:19.066929 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/stf-smoketest-smoke1-6s297"] Oct 08 00:31:19 crc kubenswrapper[4967]: I1008 00:31:19.099211 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collectd-config\" (UniqueName: \"kubernetes.io/configmap/5b0ecff1-3278-436a-9fd8-ac4ce29afe03-collectd-config\") pod \"stf-smoketest-smoke1-6s297\" (UID: \"5b0ecff1-3278-436a-9fd8-ac4ce29afe03\") " pod="service-telemetry/stf-smoketest-smoke1-6s297" Oct 08 00:31:19 crc kubenswrapper[4967]: I1008 00:31:19.099318 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/5b0ecff1-3278-436a-9fd8-ac4ce29afe03-ceilometer-entrypoint-script\") pod \"stf-smoketest-smoke1-6s297\" (UID: \"5b0ecff1-3278-436a-9fd8-ac4ce29afe03\") " pod="service-telemetry/stf-smoketest-smoke1-6s297" Oct 08 00:31:19 crc kubenswrapper[4967]: I1008 00:31:19.099548 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"healthcheck-log\" (UniqueName: \"kubernetes.io/configmap/5b0ecff1-3278-436a-9fd8-ac4ce29afe03-healthcheck-log\") pod \"stf-smoketest-smoke1-6s297\" (UID: \"5b0ecff1-3278-436a-9fd8-ac4ce29afe03\") " pod="service-telemetry/stf-smoketest-smoke1-6s297" Oct 08 00:31:19 crc kubenswrapper[4967]: I1008 00:31:19.099608 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-publisher\" (UniqueName: \"kubernetes.io/configmap/5b0ecff1-3278-436a-9fd8-ac4ce29afe03-ceilometer-publisher\") pod \"stf-smoketest-smoke1-6s297\" (UID: \"5b0ecff1-3278-436a-9fd8-ac4ce29afe03\") " pod="service-telemetry/stf-smoketest-smoke1-6s297" Oct 08 00:31:19 crc kubenswrapper[4967]: I1008 00:31:19.099733 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sensubility-config\" (UniqueName: \"kubernetes.io/configmap/5b0ecff1-3278-436a-9fd8-ac4ce29afe03-sensubility-config\") pod \"stf-smoketest-smoke1-6s297\" (UID: \"5b0ecff1-3278-436a-9fd8-ac4ce29afe03\") " pod="service-telemetry/stf-smoketest-smoke1-6s297" Oct 08 00:31:19 crc kubenswrapper[4967]: I1008 00:31:19.099797 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collectd-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/5b0ecff1-3278-436a-9fd8-ac4ce29afe03-collectd-entrypoint-script\") pod \"stf-smoketest-smoke1-6s297\" (UID: \"5b0ecff1-3278-436a-9fd8-ac4ce29afe03\") " pod="service-telemetry/stf-smoketest-smoke1-6s297" Oct 08 00:31:19 crc kubenswrapper[4967]: I1008 00:31:19.099848 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqksn\" (UniqueName: \"kubernetes.io/projected/5b0ecff1-3278-436a-9fd8-ac4ce29afe03-kube-api-access-jqksn\") pod \"stf-smoketest-smoke1-6s297\" (UID: \"5b0ecff1-3278-436a-9fd8-ac4ce29afe03\") " pod="service-telemetry/stf-smoketest-smoke1-6s297" Oct 08 00:31:19 crc kubenswrapper[4967]: I1008 00:31:19.200816 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sensubility-config\" (UniqueName: \"kubernetes.io/configmap/5b0ecff1-3278-436a-9fd8-ac4ce29afe03-sensubility-config\") pod \"stf-smoketest-smoke1-6s297\" (UID: \"5b0ecff1-3278-436a-9fd8-ac4ce29afe03\") " pod="service-telemetry/stf-smoketest-smoke1-6s297" Oct 08 00:31:19 crc kubenswrapper[4967]: I1008 00:31:19.200907 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqksn\" (UniqueName: \"kubernetes.io/projected/5b0ecff1-3278-436a-9fd8-ac4ce29afe03-kube-api-access-jqksn\") pod \"stf-smoketest-smoke1-6s297\" (UID: \"5b0ecff1-3278-436a-9fd8-ac4ce29afe03\") " pod="service-telemetry/stf-smoketest-smoke1-6s297" Oct 08 00:31:19 crc kubenswrapper[4967]: I1008 00:31:19.200945 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collectd-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/5b0ecff1-3278-436a-9fd8-ac4ce29afe03-collectd-entrypoint-script\") pod \"stf-smoketest-smoke1-6s297\" (UID: \"5b0ecff1-3278-436a-9fd8-ac4ce29afe03\") " pod="service-telemetry/stf-smoketest-smoke1-6s297" Oct 08 00:31:19 crc kubenswrapper[4967]: I1008 00:31:19.201014 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collectd-config\" (UniqueName: \"kubernetes.io/configmap/5b0ecff1-3278-436a-9fd8-ac4ce29afe03-collectd-config\") pod \"stf-smoketest-smoke1-6s297\" (UID: \"5b0ecff1-3278-436a-9fd8-ac4ce29afe03\") " pod="service-telemetry/stf-smoketest-smoke1-6s297" Oct 08 00:31:19 crc kubenswrapper[4967]: I1008 00:31:19.201078 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/5b0ecff1-3278-436a-9fd8-ac4ce29afe03-ceilometer-entrypoint-script\") pod \"stf-smoketest-smoke1-6s297\" (UID: \"5b0ecff1-3278-436a-9fd8-ac4ce29afe03\") " pod="service-telemetry/stf-smoketest-smoke1-6s297" Oct 08 00:31:19 crc kubenswrapper[4967]: I1008 00:31:19.201271 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"healthcheck-log\" (UniqueName: \"kubernetes.io/configmap/5b0ecff1-3278-436a-9fd8-ac4ce29afe03-healthcheck-log\") pod \"stf-smoketest-smoke1-6s297\" (UID: \"5b0ecff1-3278-436a-9fd8-ac4ce29afe03\") " pod="service-telemetry/stf-smoketest-smoke1-6s297" Oct 08 00:31:19 crc kubenswrapper[4967]: I1008 00:31:19.201314 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-publisher\" (UniqueName: \"kubernetes.io/configmap/5b0ecff1-3278-436a-9fd8-ac4ce29afe03-ceilometer-publisher\") pod \"stf-smoketest-smoke1-6s297\" (UID: \"5b0ecff1-3278-436a-9fd8-ac4ce29afe03\") " pod="service-telemetry/stf-smoketest-smoke1-6s297" Oct 08 00:31:19 crc kubenswrapper[4967]: I1008 00:31:19.202554 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sensubility-config\" (UniqueName: \"kubernetes.io/configmap/5b0ecff1-3278-436a-9fd8-ac4ce29afe03-sensubility-config\") pod \"stf-smoketest-smoke1-6s297\" (UID: \"5b0ecff1-3278-436a-9fd8-ac4ce29afe03\") " pod="service-telemetry/stf-smoketest-smoke1-6s297" Oct 08 00:31:19 crc kubenswrapper[4967]: I1008 00:31:19.202600 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collectd-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/5b0ecff1-3278-436a-9fd8-ac4ce29afe03-collectd-entrypoint-script\") pod \"stf-smoketest-smoke1-6s297\" (UID: \"5b0ecff1-3278-436a-9fd8-ac4ce29afe03\") " pod="service-telemetry/stf-smoketest-smoke1-6s297" Oct 08 00:31:19 crc kubenswrapper[4967]: I1008 00:31:19.202975 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-publisher\" (UniqueName: \"kubernetes.io/configmap/5b0ecff1-3278-436a-9fd8-ac4ce29afe03-ceilometer-publisher\") pod \"stf-smoketest-smoke1-6s297\" (UID: \"5b0ecff1-3278-436a-9fd8-ac4ce29afe03\") " pod="service-telemetry/stf-smoketest-smoke1-6s297" Oct 08 00:31:19 crc kubenswrapper[4967]: I1008 00:31:19.202978 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/5b0ecff1-3278-436a-9fd8-ac4ce29afe03-ceilometer-entrypoint-script\") pod \"stf-smoketest-smoke1-6s297\" (UID: \"5b0ecff1-3278-436a-9fd8-ac4ce29afe03\") " pod="service-telemetry/stf-smoketest-smoke1-6s297" Oct 08 00:31:19 crc kubenswrapper[4967]: I1008 00:31:19.203177 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collectd-config\" (UniqueName: \"kubernetes.io/configmap/5b0ecff1-3278-436a-9fd8-ac4ce29afe03-collectd-config\") pod \"stf-smoketest-smoke1-6s297\" (UID: \"5b0ecff1-3278-436a-9fd8-ac4ce29afe03\") " pod="service-telemetry/stf-smoketest-smoke1-6s297" Oct 08 00:31:19 crc kubenswrapper[4967]: I1008 00:31:19.204016 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"healthcheck-log\" (UniqueName: \"kubernetes.io/configmap/5b0ecff1-3278-436a-9fd8-ac4ce29afe03-healthcheck-log\") pod \"stf-smoketest-smoke1-6s297\" (UID: \"5b0ecff1-3278-436a-9fd8-ac4ce29afe03\") " pod="service-telemetry/stf-smoketest-smoke1-6s297" Oct 08 00:31:19 crc kubenswrapper[4967]: I1008 00:31:19.242064 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqksn\" (UniqueName: \"kubernetes.io/projected/5b0ecff1-3278-436a-9fd8-ac4ce29afe03-kube-api-access-jqksn\") pod \"stf-smoketest-smoke1-6s297\" (UID: \"5b0ecff1-3278-436a-9fd8-ac4ce29afe03\") " pod="service-telemetry/stf-smoketest-smoke1-6s297" Oct 08 00:31:19 crc kubenswrapper[4967]: I1008 00:31:19.387545 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/stf-smoketest-smoke1-6s297" Oct 08 00:31:19 crc kubenswrapper[4967]: I1008 00:31:19.622808 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/stf-smoketest-smoke1-6s297"] Oct 08 00:31:19 crc kubenswrapper[4967]: W1008 00:31:19.639090 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5b0ecff1_3278_436a_9fd8_ac4ce29afe03.slice/crio-2a0d74ecb9736652c8b7de7bf9048abd0579831bfc5963a4fedb2784e4f85457 WatchSource:0}: Error finding container 2a0d74ecb9736652c8b7de7bf9048abd0579831bfc5963a4fedb2784e4f85457: Status 404 returned error can't find the container with id 2a0d74ecb9736652c8b7de7bf9048abd0579831bfc5963a4fedb2784e4f85457 Oct 08 00:31:20 crc kubenswrapper[4967]: I1008 00:31:20.381183 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/stf-smoketest-smoke1-6s297" event={"ID":"5b0ecff1-3278-436a-9fd8-ac4ce29afe03","Type":"ContainerStarted","Data":"b9e207c85861f0aaba98a7bdc40b8d2a11a808aa6a2fc907211d69688e23006a"} Oct 08 00:31:20 crc kubenswrapper[4967]: I1008 00:31:20.381625 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/stf-smoketest-smoke1-6s297" event={"ID":"5b0ecff1-3278-436a-9fd8-ac4ce29afe03","Type":"ContainerStarted","Data":"8a2d78c5cfe6a3be18ff5b660167af56df2fc9092944e84229098b284851cf59"} Oct 08 00:31:20 crc kubenswrapper[4967]: I1008 00:31:20.381649 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/stf-smoketest-smoke1-6s297" event={"ID":"5b0ecff1-3278-436a-9fd8-ac4ce29afe03","Type":"ContainerStarted","Data":"2a0d74ecb9736652c8b7de7bf9048abd0579831bfc5963a4fedb2784e4f85457"} Oct 08 00:31:20 crc kubenswrapper[4967]: I1008 00:31:20.413501 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/stf-smoketest-smoke1-6s297" podStartSLOduration=1.41346542 podStartE2EDuration="1.41346542s" podCreationTimestamp="2025-10-08 00:31:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 00:31:20.404893848 +0000 UTC m=+1479.055428537" watchObservedRunningTime="2025-10-08 00:31:20.41346542 +0000 UTC m=+1479.064000109" Oct 08 00:31:46 crc kubenswrapper[4967]: I1008 00:31:46.699270 4967 scope.go:117] "RemoveContainer" containerID="b61d29d4dc3065525df66102951ebde9347d69c4aad384cd158f9575fa987ea8" Oct 08 00:31:46 crc kubenswrapper[4967]: I1008 00:31:46.743552 4967 scope.go:117] "RemoveContainer" containerID="ddec5842671230989b43b33ed0074978a284dec734622e318616a56c738ed951" Oct 08 00:31:52 crc kubenswrapper[4967]: I1008 00:31:52.711958 4967 generic.go:334] "Generic (PLEG): container finished" podID="5b0ecff1-3278-436a-9fd8-ac4ce29afe03" containerID="b9e207c85861f0aaba98a7bdc40b8d2a11a808aa6a2fc907211d69688e23006a" exitCode=0 Oct 08 00:31:52 crc kubenswrapper[4967]: I1008 00:31:52.712080 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/stf-smoketest-smoke1-6s297" event={"ID":"5b0ecff1-3278-436a-9fd8-ac4ce29afe03","Type":"ContainerDied","Data":"b9e207c85861f0aaba98a7bdc40b8d2a11a808aa6a2fc907211d69688e23006a"} Oct 08 00:31:52 crc kubenswrapper[4967]: I1008 00:31:52.713544 4967 scope.go:117] "RemoveContainer" containerID="b9e207c85861f0aaba98a7bdc40b8d2a11a808aa6a2fc907211d69688e23006a" Oct 08 00:31:53 crc kubenswrapper[4967]: I1008 00:31:53.725071 4967 generic.go:334] "Generic (PLEG): container finished" podID="5b0ecff1-3278-436a-9fd8-ac4ce29afe03" containerID="8a2d78c5cfe6a3be18ff5b660167af56df2fc9092944e84229098b284851cf59" exitCode=0 Oct 08 00:31:53 crc kubenswrapper[4967]: I1008 00:31:53.725142 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/stf-smoketest-smoke1-6s297" event={"ID":"5b0ecff1-3278-436a-9fd8-ac4ce29afe03","Type":"ContainerDied","Data":"8a2d78c5cfe6a3be18ff5b660167af56df2fc9092944e84229098b284851cf59"} Oct 08 00:31:55 crc kubenswrapper[4967]: I1008 00:31:55.090496 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/stf-smoketest-smoke1-6s297" Oct 08 00:31:55 crc kubenswrapper[4967]: I1008 00:31:55.253182 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"collectd-config\" (UniqueName: \"kubernetes.io/configmap/5b0ecff1-3278-436a-9fd8-ac4ce29afe03-collectd-config\") pod \"5b0ecff1-3278-436a-9fd8-ac4ce29afe03\" (UID: \"5b0ecff1-3278-436a-9fd8-ac4ce29afe03\") " Oct 08 00:31:55 crc kubenswrapper[4967]: I1008 00:31:55.253259 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"healthcheck-log\" (UniqueName: \"kubernetes.io/configmap/5b0ecff1-3278-436a-9fd8-ac4ce29afe03-healthcheck-log\") pod \"5b0ecff1-3278-436a-9fd8-ac4ce29afe03\" (UID: \"5b0ecff1-3278-436a-9fd8-ac4ce29afe03\") " Oct 08 00:31:55 crc kubenswrapper[4967]: I1008 00:31:55.253314 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-publisher\" (UniqueName: \"kubernetes.io/configmap/5b0ecff1-3278-436a-9fd8-ac4ce29afe03-ceilometer-publisher\") pod \"5b0ecff1-3278-436a-9fd8-ac4ce29afe03\" (UID: \"5b0ecff1-3278-436a-9fd8-ac4ce29afe03\") " Oct 08 00:31:55 crc kubenswrapper[4967]: I1008 00:31:55.253447 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jqksn\" (UniqueName: \"kubernetes.io/projected/5b0ecff1-3278-436a-9fd8-ac4ce29afe03-kube-api-access-jqksn\") pod \"5b0ecff1-3278-436a-9fd8-ac4ce29afe03\" (UID: \"5b0ecff1-3278-436a-9fd8-ac4ce29afe03\") " Oct 08 00:31:55 crc kubenswrapper[4967]: I1008 00:31:55.253482 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/5b0ecff1-3278-436a-9fd8-ac4ce29afe03-ceilometer-entrypoint-script\") pod \"5b0ecff1-3278-436a-9fd8-ac4ce29afe03\" (UID: \"5b0ecff1-3278-436a-9fd8-ac4ce29afe03\") " Oct 08 00:31:55 crc kubenswrapper[4967]: I1008 00:31:55.253546 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"collectd-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/5b0ecff1-3278-436a-9fd8-ac4ce29afe03-collectd-entrypoint-script\") pod \"5b0ecff1-3278-436a-9fd8-ac4ce29afe03\" (UID: \"5b0ecff1-3278-436a-9fd8-ac4ce29afe03\") " Oct 08 00:31:55 crc kubenswrapper[4967]: I1008 00:31:55.253742 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sensubility-config\" (UniqueName: \"kubernetes.io/configmap/5b0ecff1-3278-436a-9fd8-ac4ce29afe03-sensubility-config\") pod \"5b0ecff1-3278-436a-9fd8-ac4ce29afe03\" (UID: \"5b0ecff1-3278-436a-9fd8-ac4ce29afe03\") " Oct 08 00:31:55 crc kubenswrapper[4967]: I1008 00:31:55.274951 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b0ecff1-3278-436a-9fd8-ac4ce29afe03-kube-api-access-jqksn" (OuterVolumeSpecName: "kube-api-access-jqksn") pod "5b0ecff1-3278-436a-9fd8-ac4ce29afe03" (UID: "5b0ecff1-3278-436a-9fd8-ac4ce29afe03"). InnerVolumeSpecName "kube-api-access-jqksn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:31:55 crc kubenswrapper[4967]: I1008 00:31:55.285711 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5b0ecff1-3278-436a-9fd8-ac4ce29afe03-collectd-config" (OuterVolumeSpecName: "collectd-config") pod "5b0ecff1-3278-436a-9fd8-ac4ce29afe03" (UID: "5b0ecff1-3278-436a-9fd8-ac4ce29afe03"). InnerVolumeSpecName "collectd-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:31:55 crc kubenswrapper[4967]: I1008 00:31:55.286142 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5b0ecff1-3278-436a-9fd8-ac4ce29afe03-healthcheck-log" (OuterVolumeSpecName: "healthcheck-log") pod "5b0ecff1-3278-436a-9fd8-ac4ce29afe03" (UID: "5b0ecff1-3278-436a-9fd8-ac4ce29afe03"). InnerVolumeSpecName "healthcheck-log". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:31:55 crc kubenswrapper[4967]: I1008 00:31:55.288012 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5b0ecff1-3278-436a-9fd8-ac4ce29afe03-collectd-entrypoint-script" (OuterVolumeSpecName: "collectd-entrypoint-script") pod "5b0ecff1-3278-436a-9fd8-ac4ce29afe03" (UID: "5b0ecff1-3278-436a-9fd8-ac4ce29afe03"). InnerVolumeSpecName "collectd-entrypoint-script". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:31:55 crc kubenswrapper[4967]: I1008 00:31:55.294943 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5b0ecff1-3278-436a-9fd8-ac4ce29afe03-ceilometer-entrypoint-script" (OuterVolumeSpecName: "ceilometer-entrypoint-script") pod "5b0ecff1-3278-436a-9fd8-ac4ce29afe03" (UID: "5b0ecff1-3278-436a-9fd8-ac4ce29afe03"). InnerVolumeSpecName "ceilometer-entrypoint-script". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:31:55 crc kubenswrapper[4967]: I1008 00:31:55.295662 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5b0ecff1-3278-436a-9fd8-ac4ce29afe03-sensubility-config" (OuterVolumeSpecName: "sensubility-config") pod "5b0ecff1-3278-436a-9fd8-ac4ce29afe03" (UID: "5b0ecff1-3278-436a-9fd8-ac4ce29afe03"). InnerVolumeSpecName "sensubility-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:31:55 crc kubenswrapper[4967]: I1008 00:31:55.297333 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5b0ecff1-3278-436a-9fd8-ac4ce29afe03-ceilometer-publisher" (OuterVolumeSpecName: "ceilometer-publisher") pod "5b0ecff1-3278-436a-9fd8-ac4ce29afe03" (UID: "5b0ecff1-3278-436a-9fd8-ac4ce29afe03"). InnerVolumeSpecName "ceilometer-publisher". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 00:31:55 crc kubenswrapper[4967]: I1008 00:31:55.355838 4967 reconciler_common.go:293] "Volume detached for volume \"healthcheck-log\" (UniqueName: \"kubernetes.io/configmap/5b0ecff1-3278-436a-9fd8-ac4ce29afe03-healthcheck-log\") on node \"crc\" DevicePath \"\"" Oct 08 00:31:55 crc kubenswrapper[4967]: I1008 00:31:55.355910 4967 reconciler_common.go:293] "Volume detached for volume \"ceilometer-publisher\" (UniqueName: \"kubernetes.io/configmap/5b0ecff1-3278-436a-9fd8-ac4ce29afe03-ceilometer-publisher\") on node \"crc\" DevicePath \"\"" Oct 08 00:31:55 crc kubenswrapper[4967]: I1008 00:31:55.355943 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jqksn\" (UniqueName: \"kubernetes.io/projected/5b0ecff1-3278-436a-9fd8-ac4ce29afe03-kube-api-access-jqksn\") on node \"crc\" DevicePath \"\"" Oct 08 00:31:55 crc kubenswrapper[4967]: I1008 00:31:55.355969 4967 reconciler_common.go:293] "Volume detached for volume \"ceilometer-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/5b0ecff1-3278-436a-9fd8-ac4ce29afe03-ceilometer-entrypoint-script\") on node \"crc\" DevicePath \"\"" Oct 08 00:31:55 crc kubenswrapper[4967]: I1008 00:31:55.355995 4967 reconciler_common.go:293] "Volume detached for volume \"collectd-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/5b0ecff1-3278-436a-9fd8-ac4ce29afe03-collectd-entrypoint-script\") on node \"crc\" DevicePath \"\"" Oct 08 00:31:55 crc kubenswrapper[4967]: I1008 00:31:55.356024 4967 reconciler_common.go:293] "Volume detached for volume \"sensubility-config\" (UniqueName: \"kubernetes.io/configmap/5b0ecff1-3278-436a-9fd8-ac4ce29afe03-sensubility-config\") on node \"crc\" DevicePath \"\"" Oct 08 00:31:55 crc kubenswrapper[4967]: I1008 00:31:55.356046 4967 reconciler_common.go:293] "Volume detached for volume \"collectd-config\" (UniqueName: \"kubernetes.io/configmap/5b0ecff1-3278-436a-9fd8-ac4ce29afe03-collectd-config\") on node \"crc\" DevicePath \"\"" Oct 08 00:31:55 crc kubenswrapper[4967]: I1008 00:31:55.745176 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/stf-smoketest-smoke1-6s297" event={"ID":"5b0ecff1-3278-436a-9fd8-ac4ce29afe03","Type":"ContainerDied","Data":"2a0d74ecb9736652c8b7de7bf9048abd0579831bfc5963a4fedb2784e4f85457"} Oct 08 00:31:55 crc kubenswrapper[4967]: I1008 00:31:55.745213 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2a0d74ecb9736652c8b7de7bf9048abd0579831bfc5963a4fedb2784e4f85457" Oct 08 00:31:55 crc kubenswrapper[4967]: I1008 00:31:55.745537 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/stf-smoketest-smoke1-6s297" Oct 08 00:31:57 crc kubenswrapper[4967]: I1008 00:31:57.396191 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_stf-smoketest-smoke1-6s297_5b0ecff1-3278-436a-9fd8-ac4ce29afe03/smoketest-collectd/0.log" Oct 08 00:31:57 crc kubenswrapper[4967]: I1008 00:31:57.706736 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_stf-smoketest-smoke1-6s297_5b0ecff1-3278-436a-9fd8-ac4ce29afe03/smoketest-ceilometer/0.log" Oct 08 00:31:58 crc kubenswrapper[4967]: I1008 00:31:58.061390 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-interconnect-68864d46cb-nb9sh_94d8be0a-4986-4ff3-b327-b6da51cfdd64/default-interconnect/0.log" Oct 08 00:31:58 crc kubenswrapper[4967]: I1008 00:31:58.409180 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-coll-meter-smartgateway-7cd87f9766-qcv8n_ddba1803-397c-4daf-bc56-4857967229d1/bridge/2.log" Oct 08 00:31:58 crc kubenswrapper[4967]: I1008 00:31:58.786625 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-coll-meter-smartgateway-7cd87f9766-qcv8n_ddba1803-397c-4daf-bc56-4857967229d1/sg-core/0.log" Oct 08 00:31:59 crc kubenswrapper[4967]: I1008 00:31:59.115106 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-coll-event-smartgateway-6f7464cc9d-bjhng_84dadc03-b18c-490c-adb2-9d3f605865d1/bridge/2.log" Oct 08 00:31:59 crc kubenswrapper[4967]: I1008 00:31:59.369724 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-coll-event-smartgateway-6f7464cc9d-bjhng_84dadc03-b18c-490c-adb2-9d3f605865d1/sg-core/0.log" Oct 08 00:31:59 crc kubenswrapper[4967]: I1008 00:31:59.726043 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-ceil-meter-smartgateway-57948895dc-dlql4_1f1c83bb-ff16-45ff-8fe5-e1b7e9d5166b/bridge/2.log" Oct 08 00:32:00 crc kubenswrapper[4967]: I1008 00:32:00.008648 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-ceil-meter-smartgateway-57948895dc-dlql4_1f1c83bb-ff16-45ff-8fe5-e1b7e9d5166b/sg-core/0.log" Oct 08 00:32:00 crc kubenswrapper[4967]: I1008 00:32:00.305482 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-ceil-event-smartgateway-5d47c9955c-t24dp_b3e181c2-79f9-47be-b4fd-f9f05599395f/bridge/2.log" Oct 08 00:32:00 crc kubenswrapper[4967]: I1008 00:32:00.618812 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-ceil-event-smartgateway-5d47c9955c-t24dp_b3e181c2-79f9-47be-b4fd-f9f05599395f/sg-core/0.log" Oct 08 00:32:00 crc kubenswrapper[4967]: I1008 00:32:00.929246 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-sens-meter-smartgateway-5759b4d97-rwnt8_a9da825c-48d2-477d-b8c5-1ac878239c37/bridge/2.log" Oct 08 00:32:01 crc kubenswrapper[4967]: I1008 00:32:01.236323 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-cloud1-sens-meter-smartgateway-5759b4d97-rwnt8_a9da825c-48d2-477d-b8c5-1ac878239c37/sg-core/0.log" Oct 08 00:32:04 crc kubenswrapper[4967]: I1008 00:32:04.878912 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_smart-gateway-operator-7d98fcf998-q8f6v_5bc8d2b2-4afa-44de-a704-abacdd934b28/operator/0.log" Oct 08 00:32:05 crc kubenswrapper[4967]: I1008 00:32:05.223993 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_prometheus-default-0_82ce08ad-02fd-4ae6-b3c5-c2c01ac6fdec/prometheus/0.log" Oct 08 00:32:05 crc kubenswrapper[4967]: I1008 00:32:05.569388 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_elasticsearch-es-default-0_79b69a1c-270d-4261-b80f-3aff7be1c7bf/elasticsearch/0.log" Oct 08 00:32:05 crc kubenswrapper[4967]: I1008 00:32:05.865811 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_default-snmp-webhook-6856cfb745-6jgkj_f3eb705e-460b-45f4-be46-8d811bda63c3/prometheus-webhook-snmp/0.log" Oct 08 00:32:06 crc kubenswrapper[4967]: I1008 00:32:06.223849 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_alertmanager-default-0_9ce411af-8da3-4002-a75b-bc8ba4a64db8/alertmanager/0.log" Oct 08 00:32:21 crc kubenswrapper[4967]: I1008 00:32:21.472969 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-56dc879589-jp9r2_5f3f5df9-b819-464d-b7ca-2dea8ceab64d/operator/0.log" Oct 08 00:32:25 crc kubenswrapper[4967]: I1008 00:32:25.078636 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_smart-gateway-operator-7d98fcf998-q8f6v_5bc8d2b2-4afa-44de-a704-abacdd934b28/operator/0.log" Oct 08 00:32:25 crc kubenswrapper[4967]: I1008 00:32:25.376420 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_qdr-test_c8219ce4-173e-47ef-810d-6dc298346972/qdr/0.log" Oct 08 00:33:04 crc kubenswrapper[4967]: I1008 00:33:04.298876 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-nht2s/must-gather-8bzm4"] Oct 08 00:33:04 crc kubenswrapper[4967]: E1008 00:33:04.299768 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b0ecff1-3278-436a-9fd8-ac4ce29afe03" containerName="smoketest-ceilometer" Oct 08 00:33:04 crc kubenswrapper[4967]: I1008 00:33:04.299787 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b0ecff1-3278-436a-9fd8-ac4ce29afe03" containerName="smoketest-ceilometer" Oct 08 00:33:04 crc kubenswrapper[4967]: E1008 00:33:04.299809 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b0ecff1-3278-436a-9fd8-ac4ce29afe03" containerName="smoketest-collectd" Oct 08 00:33:04 crc kubenswrapper[4967]: I1008 00:33:04.299822 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b0ecff1-3278-436a-9fd8-ac4ce29afe03" containerName="smoketest-collectd" Oct 08 00:33:04 crc kubenswrapper[4967]: I1008 00:33:04.300003 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b0ecff1-3278-436a-9fd8-ac4ce29afe03" containerName="smoketest-ceilometer" Oct 08 00:33:04 crc kubenswrapper[4967]: I1008 00:33:04.300030 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b0ecff1-3278-436a-9fd8-ac4ce29afe03" containerName="smoketest-collectd" Oct 08 00:33:04 crc kubenswrapper[4967]: I1008 00:33:04.300876 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-nht2s/must-gather-8bzm4" Oct 08 00:33:04 crc kubenswrapper[4967]: I1008 00:33:04.303730 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-nht2s"/"kube-root-ca.crt" Oct 08 00:33:04 crc kubenswrapper[4967]: I1008 00:33:04.309013 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-nht2s"/"openshift-service-ca.crt" Oct 08 00:33:04 crc kubenswrapper[4967]: I1008 00:33:04.338123 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-nht2s/must-gather-8bzm4"] Oct 08 00:33:04 crc kubenswrapper[4967]: I1008 00:33:04.439661 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zwkdr\" (UniqueName: \"kubernetes.io/projected/2b4b9b9e-7a8f-43ef-a386-6afc3fe66b0c-kube-api-access-zwkdr\") pod \"must-gather-8bzm4\" (UID: \"2b4b9b9e-7a8f-43ef-a386-6afc3fe66b0c\") " pod="openshift-must-gather-nht2s/must-gather-8bzm4" Oct 08 00:33:04 crc kubenswrapper[4967]: I1008 00:33:04.439740 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2b4b9b9e-7a8f-43ef-a386-6afc3fe66b0c-must-gather-output\") pod \"must-gather-8bzm4\" (UID: \"2b4b9b9e-7a8f-43ef-a386-6afc3fe66b0c\") " pod="openshift-must-gather-nht2s/must-gather-8bzm4" Oct 08 00:33:04 crc kubenswrapper[4967]: I1008 00:33:04.541149 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zwkdr\" (UniqueName: \"kubernetes.io/projected/2b4b9b9e-7a8f-43ef-a386-6afc3fe66b0c-kube-api-access-zwkdr\") pod \"must-gather-8bzm4\" (UID: \"2b4b9b9e-7a8f-43ef-a386-6afc3fe66b0c\") " pod="openshift-must-gather-nht2s/must-gather-8bzm4" Oct 08 00:33:04 crc kubenswrapper[4967]: I1008 00:33:04.541197 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2b4b9b9e-7a8f-43ef-a386-6afc3fe66b0c-must-gather-output\") pod \"must-gather-8bzm4\" (UID: \"2b4b9b9e-7a8f-43ef-a386-6afc3fe66b0c\") " pod="openshift-must-gather-nht2s/must-gather-8bzm4" Oct 08 00:33:04 crc kubenswrapper[4967]: I1008 00:33:04.541563 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2b4b9b9e-7a8f-43ef-a386-6afc3fe66b0c-must-gather-output\") pod \"must-gather-8bzm4\" (UID: \"2b4b9b9e-7a8f-43ef-a386-6afc3fe66b0c\") " pod="openshift-must-gather-nht2s/must-gather-8bzm4" Oct 08 00:33:04 crc kubenswrapper[4967]: I1008 00:33:04.572320 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zwkdr\" (UniqueName: \"kubernetes.io/projected/2b4b9b9e-7a8f-43ef-a386-6afc3fe66b0c-kube-api-access-zwkdr\") pod \"must-gather-8bzm4\" (UID: \"2b4b9b9e-7a8f-43ef-a386-6afc3fe66b0c\") " pod="openshift-must-gather-nht2s/must-gather-8bzm4" Oct 08 00:33:04 crc kubenswrapper[4967]: I1008 00:33:04.624378 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-nht2s/must-gather-8bzm4" Oct 08 00:33:05 crc kubenswrapper[4967]: I1008 00:33:05.079849 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-nht2s/must-gather-8bzm4"] Oct 08 00:33:05 crc kubenswrapper[4967]: I1008 00:33:05.378305 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-nht2s/must-gather-8bzm4" event={"ID":"2b4b9b9e-7a8f-43ef-a386-6afc3fe66b0c","Type":"ContainerStarted","Data":"1c2daeb25f40abafde6e4e1865ff9e99ff1078fc03f56a9061c382cfd11754ba"} Oct 08 00:33:09 crc kubenswrapper[4967]: I1008 00:33:09.408104 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-nht2s/must-gather-8bzm4" event={"ID":"2b4b9b9e-7a8f-43ef-a386-6afc3fe66b0c","Type":"ContainerStarted","Data":"8af31072c4b2e54caa14002d88e4b2fe42487321e251f177934d28e3d905c948"} Oct 08 00:33:09 crc kubenswrapper[4967]: I1008 00:33:09.408730 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-nht2s/must-gather-8bzm4" event={"ID":"2b4b9b9e-7a8f-43ef-a386-6afc3fe66b0c","Type":"ContainerStarted","Data":"21d17dc5b2a6acf21805f7da541ef94fb231de8098d35b6e2f1304f68d614631"} Oct 08 00:33:09 crc kubenswrapper[4967]: I1008 00:33:09.428103 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-nht2s/must-gather-8bzm4" podStartSLOduration=1.604840243 podStartE2EDuration="5.42808663s" podCreationTimestamp="2025-10-08 00:33:04 +0000 UTC" firstStartedPulling="2025-10-08 00:33:05.094773994 +0000 UTC m=+1583.745308643" lastFinishedPulling="2025-10-08 00:33:08.918020351 +0000 UTC m=+1587.568555030" observedRunningTime="2025-10-08 00:33:09.424540014 +0000 UTC m=+1588.075074673" watchObservedRunningTime="2025-10-08 00:33:09.42808663 +0000 UTC m=+1588.078621269" Oct 08 00:33:36 crc kubenswrapper[4967]: I1008 00:33:36.920254 4967 patch_prober.go:28] interesting pod/machine-config-daemon-xvdtp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 00:33:36 crc kubenswrapper[4967]: I1008 00:33:36.920905 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" podUID="9461162a-c27d-465e-9ca7-2d3936877cd8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 00:33:53 crc kubenswrapper[4967]: I1008 00:33:53.439419 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-lpxcq_cd33606b-2d0f-4f4b-bc86-b85ad4f56f83/control-plane-machine-set-operator/0.log" Oct 08 00:33:53 crc kubenswrapper[4967]: I1008 00:33:53.497861 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-ljjfc_1a14c5f6-9dfd-47fd-acf7-65ec6bbd5149/kube-rbac-proxy/0.log" Oct 08 00:33:53 crc kubenswrapper[4967]: I1008 00:33:53.635290 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-ljjfc_1a14c5f6-9dfd-47fd-acf7-65ec6bbd5149/machine-api-operator/0.log" Oct 08 00:34:05 crc kubenswrapper[4967]: I1008 00:34:05.450526 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-7d4cc89fcb-fftc7_59077174-55c7-4ca3-8ad4-9252240ead8c/cert-manager-controller/0.log" Oct 08 00:34:05 crc kubenswrapper[4967]: I1008 00:34:05.652761 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-d969966f-mr5n2_e13c83b9-8e80-4bf3-987d-a2b8de2de7be/cert-manager-webhook/0.log" Oct 08 00:34:05 crc kubenswrapper[4967]: I1008 00:34:05.657569 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7d9f95dbf-6r7sw_8b29e193-8366-4f9e-b3c8-761aae502121/cert-manager-cainjector/0.log" Oct 08 00:34:06 crc kubenswrapper[4967]: I1008 00:34:06.920568 4967 patch_prober.go:28] interesting pod/machine-config-daemon-xvdtp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 00:34:06 crc kubenswrapper[4967]: I1008 00:34:06.921009 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" podUID="9461162a-c27d-465e-9ca7-2d3936877cd8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 00:34:21 crc kubenswrapper[4967]: I1008 00:34:21.195038 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69dw25g_277b8c1b-6e5d-44f2-b17b-594c3b367dab/util/0.log" Oct 08 00:34:21 crc kubenswrapper[4967]: I1008 00:34:21.394173 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69dw25g_277b8c1b-6e5d-44f2-b17b-594c3b367dab/util/0.log" Oct 08 00:34:21 crc kubenswrapper[4967]: I1008 00:34:21.425210 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69dw25g_277b8c1b-6e5d-44f2-b17b-594c3b367dab/pull/0.log" Oct 08 00:34:21 crc kubenswrapper[4967]: I1008 00:34:21.446967 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69dw25g_277b8c1b-6e5d-44f2-b17b-594c3b367dab/pull/0.log" Oct 08 00:34:21 crc kubenswrapper[4967]: I1008 00:34:21.580778 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69dw25g_277b8c1b-6e5d-44f2-b17b-594c3b367dab/util/0.log" Oct 08 00:34:21 crc kubenswrapper[4967]: I1008 00:34:21.599015 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69dw25g_277b8c1b-6e5d-44f2-b17b-594c3b367dab/pull/0.log" Oct 08 00:34:21 crc kubenswrapper[4967]: I1008 00:34:21.620273 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69dw25g_277b8c1b-6e5d-44f2-b17b-594c3b367dab/extract/0.log" Oct 08 00:34:21 crc kubenswrapper[4967]: I1008 00:34:21.781429 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fd24kq_3221812d-48b9-4361-a75e-3472b715d548/util/0.log" Oct 08 00:34:21 crc kubenswrapper[4967]: I1008 00:34:21.956171 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fd24kq_3221812d-48b9-4361-a75e-3472b715d548/util/0.log" Oct 08 00:34:21 crc kubenswrapper[4967]: I1008 00:34:21.967480 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fd24kq_3221812d-48b9-4361-a75e-3472b715d548/pull/0.log" Oct 08 00:34:21 crc kubenswrapper[4967]: I1008 00:34:21.987857 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fd24kq_3221812d-48b9-4361-a75e-3472b715d548/pull/0.log" Oct 08 00:34:22 crc kubenswrapper[4967]: I1008 00:34:22.157940 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fd24kq_3221812d-48b9-4361-a75e-3472b715d548/extract/0.log" Oct 08 00:34:22 crc kubenswrapper[4967]: I1008 00:34:22.165081 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fd24kq_3221812d-48b9-4361-a75e-3472b715d548/pull/0.log" Oct 08 00:34:22 crc kubenswrapper[4967]: I1008 00:34:22.179525 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fd24kq_3221812d-48b9-4361-a75e-3472b715d548/util/0.log" Oct 08 00:34:22 crc kubenswrapper[4967]: I1008 00:34:22.336836 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d4n2bh_55341c94-8af6-4a61-8fa8-913ced0e2732/util/0.log" Oct 08 00:34:22 crc kubenswrapper[4967]: I1008 00:34:22.488383 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d4n2bh_55341c94-8af6-4a61-8fa8-913ced0e2732/util/0.log" Oct 08 00:34:22 crc kubenswrapper[4967]: I1008 00:34:22.492316 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d4n2bh_55341c94-8af6-4a61-8fa8-913ced0e2732/pull/0.log" Oct 08 00:34:22 crc kubenswrapper[4967]: I1008 00:34:22.543183 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d4n2bh_55341c94-8af6-4a61-8fa8-913ced0e2732/pull/0.log" Oct 08 00:34:22 crc kubenswrapper[4967]: I1008 00:34:22.687035 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d4n2bh_55341c94-8af6-4a61-8fa8-913ced0e2732/pull/0.log" Oct 08 00:34:22 crc kubenswrapper[4967]: I1008 00:34:22.691515 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d4n2bh_55341c94-8af6-4a61-8fa8-913ced0e2732/util/0.log" Oct 08 00:34:22 crc kubenswrapper[4967]: I1008 00:34:22.742703 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d4n2bh_55341c94-8af6-4a61-8fa8-913ced0e2732/extract/0.log" Oct 08 00:34:22 crc kubenswrapper[4967]: I1008 00:34:22.862000 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-hk9rd_7355878e-878d-494a-98f3-f4a8f6c0c613/extract-utilities/0.log" Oct 08 00:34:23 crc kubenswrapper[4967]: I1008 00:34:23.044037 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-hk9rd_7355878e-878d-494a-98f3-f4a8f6c0c613/extract-utilities/0.log" Oct 08 00:34:23 crc kubenswrapper[4967]: I1008 00:34:23.047171 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-hk9rd_7355878e-878d-494a-98f3-f4a8f6c0c613/extract-content/0.log" Oct 08 00:34:23 crc kubenswrapper[4967]: I1008 00:34:23.047254 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-hk9rd_7355878e-878d-494a-98f3-f4a8f6c0c613/extract-content/0.log" Oct 08 00:34:23 crc kubenswrapper[4967]: I1008 00:34:23.195171 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-hk9rd_7355878e-878d-494a-98f3-f4a8f6c0c613/extract-utilities/0.log" Oct 08 00:34:23 crc kubenswrapper[4967]: I1008 00:34:23.240182 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-hk9rd_7355878e-878d-494a-98f3-f4a8f6c0c613/extract-content/0.log" Oct 08 00:34:23 crc kubenswrapper[4967]: I1008 00:34:23.407564 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-wc5gb_36c8543a-dbe6-4aac-bb98-f39558beb387/extract-utilities/0.log" Oct 08 00:34:23 crc kubenswrapper[4967]: I1008 00:34:23.496411 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-hk9rd_7355878e-878d-494a-98f3-f4a8f6c0c613/registry-server/0.log" Oct 08 00:34:23 crc kubenswrapper[4967]: I1008 00:34:23.572768 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-wc5gb_36c8543a-dbe6-4aac-bb98-f39558beb387/extract-content/0.log" Oct 08 00:34:23 crc kubenswrapper[4967]: I1008 00:34:23.578627 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-wc5gb_36c8543a-dbe6-4aac-bb98-f39558beb387/extract-utilities/0.log" Oct 08 00:34:23 crc kubenswrapper[4967]: I1008 00:34:23.579050 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-wc5gb_36c8543a-dbe6-4aac-bb98-f39558beb387/extract-content/0.log" Oct 08 00:34:23 crc kubenswrapper[4967]: I1008 00:34:23.724881 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-wc5gb_36c8543a-dbe6-4aac-bb98-f39558beb387/extract-utilities/0.log" Oct 08 00:34:23 crc kubenswrapper[4967]: I1008 00:34:23.749866 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-wc5gb_36c8543a-dbe6-4aac-bb98-f39558beb387/extract-content/0.log" Oct 08 00:34:23 crc kubenswrapper[4967]: I1008 00:34:23.914438 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_dec6855936450d9745a6250315030408d50c8ea57f37cefa88be37c927pvlxl_330d296a-13cf-4883-8d5b-af3896f592b8/util/0.log" Oct 08 00:34:24 crc kubenswrapper[4967]: I1008 00:34:24.168250 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_dec6855936450d9745a6250315030408d50c8ea57f37cefa88be37c927pvlxl_330d296a-13cf-4883-8d5b-af3896f592b8/pull/0.log" Oct 08 00:34:24 crc kubenswrapper[4967]: I1008 00:34:24.168440 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_dec6855936450d9745a6250315030408d50c8ea57f37cefa88be37c927pvlxl_330d296a-13cf-4883-8d5b-af3896f592b8/pull/0.log" Oct 08 00:34:24 crc kubenswrapper[4967]: I1008 00:34:24.183039 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_dec6855936450d9745a6250315030408d50c8ea57f37cefa88be37c927pvlxl_330d296a-13cf-4883-8d5b-af3896f592b8/util/0.log" Oct 08 00:34:24 crc kubenswrapper[4967]: I1008 00:34:24.183101 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-wc5gb_36c8543a-dbe6-4aac-bb98-f39558beb387/registry-server/0.log" Oct 08 00:34:24 crc kubenswrapper[4967]: I1008 00:34:24.335722 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_dec6855936450d9745a6250315030408d50c8ea57f37cefa88be37c927pvlxl_330d296a-13cf-4883-8d5b-af3896f592b8/util/0.log" Oct 08 00:34:24 crc kubenswrapper[4967]: I1008 00:34:24.378130 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_dec6855936450d9745a6250315030408d50c8ea57f37cefa88be37c927pvlxl_330d296a-13cf-4883-8d5b-af3896f592b8/pull/0.log" Oct 08 00:34:24 crc kubenswrapper[4967]: I1008 00:34:24.378366 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_dec6855936450d9745a6250315030408d50c8ea57f37cefa88be37c927pvlxl_330d296a-13cf-4883-8d5b-af3896f592b8/extract/0.log" Oct 08 00:34:24 crc kubenswrapper[4967]: I1008 00:34:24.387817 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-szxjs_6acb8d28-53a4-46b1-93ef-5c0666234019/marketplace-operator/0.log" Oct 08 00:34:24 crc kubenswrapper[4967]: I1008 00:34:24.527618 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-6gq5j_17ee9c06-c101-414a-ad72-753c071c0cda/extract-utilities/0.log" Oct 08 00:34:24 crc kubenswrapper[4967]: I1008 00:34:24.700339 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-6gq5j_17ee9c06-c101-414a-ad72-753c071c0cda/extract-content/0.log" Oct 08 00:34:24 crc kubenswrapper[4967]: I1008 00:34:24.719036 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-6gq5j_17ee9c06-c101-414a-ad72-753c071c0cda/extract-utilities/0.log" Oct 08 00:34:24 crc kubenswrapper[4967]: I1008 00:34:24.744332 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-6gq5j_17ee9c06-c101-414a-ad72-753c071c0cda/extract-content/0.log" Oct 08 00:34:24 crc kubenswrapper[4967]: I1008 00:34:24.890278 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-6gq5j_17ee9c06-c101-414a-ad72-753c071c0cda/extract-content/0.log" Oct 08 00:34:24 crc kubenswrapper[4967]: I1008 00:34:24.893088 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-6gq5j_17ee9c06-c101-414a-ad72-753c071c0cda/extract-utilities/0.log" Oct 08 00:34:25 crc kubenswrapper[4967]: I1008 00:34:25.213994 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-6gq5j_17ee9c06-c101-414a-ad72-753c071c0cda/registry-server/0.log" Oct 08 00:34:36 crc kubenswrapper[4967]: I1008 00:34:36.920052 4967 patch_prober.go:28] interesting pod/machine-config-daemon-xvdtp container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 00:34:36 crc kubenswrapper[4967]: I1008 00:34:36.920656 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" podUID="9461162a-c27d-465e-9ca7-2d3936877cd8" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 00:34:36 crc kubenswrapper[4967]: I1008 00:34:36.920734 4967 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" Oct 08 00:34:36 crc kubenswrapper[4967]: I1008 00:34:36.921450 4967 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"dd4b0abe7436dc2e0c64a27b31f4a959aef1a772ba57812e2a8fbddc81d33d75"} pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 00:34:36 crc kubenswrapper[4967]: I1008 00:34:36.921511 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" podUID="9461162a-c27d-465e-9ca7-2d3936877cd8" containerName="machine-config-daemon" containerID="cri-o://dd4b0abe7436dc2e0c64a27b31f4a959aef1a772ba57812e2a8fbddc81d33d75" gracePeriod=600 Oct 08 00:34:37 crc kubenswrapper[4967]: E1008 00:34:37.062303 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xvdtp_openshift-machine-config-operator(9461162a-c27d-465e-9ca7-2d3936877cd8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" podUID="9461162a-c27d-465e-9ca7-2d3936877cd8" Oct 08 00:34:37 crc kubenswrapper[4967]: I1008 00:34:37.126321 4967 generic.go:334] "Generic (PLEG): container finished" podID="9461162a-c27d-465e-9ca7-2d3936877cd8" containerID="dd4b0abe7436dc2e0c64a27b31f4a959aef1a772ba57812e2a8fbddc81d33d75" exitCode=0 Oct 08 00:34:37 crc kubenswrapper[4967]: I1008 00:34:37.126618 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" event={"ID":"9461162a-c27d-465e-9ca7-2d3936877cd8","Type":"ContainerDied","Data":"dd4b0abe7436dc2e0c64a27b31f4a959aef1a772ba57812e2a8fbddc81d33d75"} Oct 08 00:34:37 crc kubenswrapper[4967]: I1008 00:34:37.126779 4967 scope.go:117] "RemoveContainer" containerID="5e1e0751dcae2b0e4bad8c7f2220f3bca21a69c9cccc3dd960fa9770bb1adf17" Oct 08 00:34:37 crc kubenswrapper[4967]: I1008 00:34:37.127432 4967 scope.go:117] "RemoveContainer" containerID="dd4b0abe7436dc2e0c64a27b31f4a959aef1a772ba57812e2a8fbddc81d33d75" Oct 08 00:34:37 crc kubenswrapper[4967]: E1008 00:34:37.127806 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xvdtp_openshift-machine-config-operator(9461162a-c27d-465e-9ca7-2d3936877cd8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" podUID="9461162a-c27d-465e-9ca7-2d3936877cd8" Oct 08 00:34:37 crc kubenswrapper[4967]: I1008 00:34:37.401030 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-7c8cf85677-88lmm_fb355e35-7b9c-4c0d-bb62-aa169f2bef5a/prometheus-operator/0.log" Oct 08 00:34:37 crc kubenswrapper[4967]: I1008 00:34:37.527079 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-5fc4456945-kpznr_2a5fd0cb-9c8e-41e9-af18-d18ddf30180e/prometheus-operator-admission-webhook/0.log" Oct 08 00:34:37 crc kubenswrapper[4967]: I1008 00:34:37.565986 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-5fc4456945-q69j8_a57260c5-8756-489a-be29-6e71c74778bc/prometheus-operator-admission-webhook/0.log" Oct 08 00:34:37 crc kubenswrapper[4967]: I1008 00:34:37.699135 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-cc5f78dfc-sgk4j_d1a621b2-6929-40de-97b1-01a020d70aa2/operator/0.log" Oct 08 00:34:37 crc kubenswrapper[4967]: I1008 00:34:37.772994 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-54bc95c9fb-x476n_478b5edf-299e-4d19-bc8f-02a5e9774148/perses-operator/0.log" Oct 08 00:34:52 crc kubenswrapper[4967]: I1008 00:34:52.126234 4967 scope.go:117] "RemoveContainer" containerID="dd4b0abe7436dc2e0c64a27b31f4a959aef1a772ba57812e2a8fbddc81d33d75" Oct 08 00:34:52 crc kubenswrapper[4967]: E1008 00:34:52.127142 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xvdtp_openshift-machine-config-operator(9461162a-c27d-465e-9ca7-2d3936877cd8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" podUID="9461162a-c27d-465e-9ca7-2d3936877cd8" Oct 08 00:35:06 crc kubenswrapper[4967]: I1008 00:35:06.124740 4967 scope.go:117] "RemoveContainer" containerID="dd4b0abe7436dc2e0c64a27b31f4a959aef1a772ba57812e2a8fbddc81d33d75" Oct 08 00:35:06 crc kubenswrapper[4967]: E1008 00:35:06.125911 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xvdtp_openshift-machine-config-operator(9461162a-c27d-465e-9ca7-2d3936877cd8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" podUID="9461162a-c27d-465e-9ca7-2d3936877cd8" Oct 08 00:35:19 crc kubenswrapper[4967]: I1008 00:35:19.564527 4967 generic.go:334] "Generic (PLEG): container finished" podID="2b4b9b9e-7a8f-43ef-a386-6afc3fe66b0c" containerID="21d17dc5b2a6acf21805f7da541ef94fb231de8098d35b6e2f1304f68d614631" exitCode=0 Oct 08 00:35:19 crc kubenswrapper[4967]: I1008 00:35:19.564636 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-nht2s/must-gather-8bzm4" event={"ID":"2b4b9b9e-7a8f-43ef-a386-6afc3fe66b0c","Type":"ContainerDied","Data":"21d17dc5b2a6acf21805f7da541ef94fb231de8098d35b6e2f1304f68d614631"} Oct 08 00:35:19 crc kubenswrapper[4967]: I1008 00:35:19.566100 4967 scope.go:117] "RemoveContainer" containerID="21d17dc5b2a6acf21805f7da541ef94fb231de8098d35b6e2f1304f68d614631" Oct 08 00:35:19 crc kubenswrapper[4967]: I1008 00:35:19.803028 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-nht2s_must-gather-8bzm4_2b4b9b9e-7a8f-43ef-a386-6afc3fe66b0c/gather/0.log" Oct 08 00:35:21 crc kubenswrapper[4967]: I1008 00:35:21.118358 4967 scope.go:117] "RemoveContainer" containerID="dd4b0abe7436dc2e0c64a27b31f4a959aef1a772ba57812e2a8fbddc81d33d75" Oct 08 00:35:21 crc kubenswrapper[4967]: E1008 00:35:21.118658 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xvdtp_openshift-machine-config-operator(9461162a-c27d-465e-9ca7-2d3936877cd8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" podUID="9461162a-c27d-465e-9ca7-2d3936877cd8" Oct 08 00:35:26 crc kubenswrapper[4967]: I1008 00:35:26.233520 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-nht2s/must-gather-8bzm4"] Oct 08 00:35:26 crc kubenswrapper[4967]: I1008 00:35:26.234307 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-nht2s/must-gather-8bzm4" podUID="2b4b9b9e-7a8f-43ef-a386-6afc3fe66b0c" containerName="copy" containerID="cri-o://8af31072c4b2e54caa14002d88e4b2fe42487321e251f177934d28e3d905c948" gracePeriod=2 Oct 08 00:35:26 crc kubenswrapper[4967]: I1008 00:35:26.246318 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-nht2s/must-gather-8bzm4"] Oct 08 00:35:26 crc kubenswrapper[4967]: I1008 00:35:26.623518 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-nht2s_must-gather-8bzm4_2b4b9b9e-7a8f-43ef-a386-6afc3fe66b0c/copy/0.log" Oct 08 00:35:26 crc kubenswrapper[4967]: I1008 00:35:26.624333 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-nht2s/must-gather-8bzm4" Oct 08 00:35:26 crc kubenswrapper[4967]: I1008 00:35:26.640623 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-nht2s_must-gather-8bzm4_2b4b9b9e-7a8f-43ef-a386-6afc3fe66b0c/copy/0.log" Oct 08 00:35:26 crc kubenswrapper[4967]: I1008 00:35:26.641859 4967 generic.go:334] "Generic (PLEG): container finished" podID="2b4b9b9e-7a8f-43ef-a386-6afc3fe66b0c" containerID="8af31072c4b2e54caa14002d88e4b2fe42487321e251f177934d28e3d905c948" exitCode=143 Oct 08 00:35:26 crc kubenswrapper[4967]: I1008 00:35:26.641949 4967 scope.go:117] "RemoveContainer" containerID="8af31072c4b2e54caa14002d88e4b2fe42487321e251f177934d28e3d905c948" Oct 08 00:35:26 crc kubenswrapper[4967]: I1008 00:35:26.641965 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-nht2s/must-gather-8bzm4" Oct 08 00:35:26 crc kubenswrapper[4967]: I1008 00:35:26.665071 4967 scope.go:117] "RemoveContainer" containerID="21d17dc5b2a6acf21805f7da541ef94fb231de8098d35b6e2f1304f68d614631" Oct 08 00:35:26 crc kubenswrapper[4967]: I1008 00:35:26.712410 4967 scope.go:117] "RemoveContainer" containerID="8af31072c4b2e54caa14002d88e4b2fe42487321e251f177934d28e3d905c948" Oct 08 00:35:26 crc kubenswrapper[4967]: E1008 00:35:26.713386 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8af31072c4b2e54caa14002d88e4b2fe42487321e251f177934d28e3d905c948\": container with ID starting with 8af31072c4b2e54caa14002d88e4b2fe42487321e251f177934d28e3d905c948 not found: ID does not exist" containerID="8af31072c4b2e54caa14002d88e4b2fe42487321e251f177934d28e3d905c948" Oct 08 00:35:26 crc kubenswrapper[4967]: I1008 00:35:26.713447 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8af31072c4b2e54caa14002d88e4b2fe42487321e251f177934d28e3d905c948"} err="failed to get container status \"8af31072c4b2e54caa14002d88e4b2fe42487321e251f177934d28e3d905c948\": rpc error: code = NotFound desc = could not find container \"8af31072c4b2e54caa14002d88e4b2fe42487321e251f177934d28e3d905c948\": container with ID starting with 8af31072c4b2e54caa14002d88e4b2fe42487321e251f177934d28e3d905c948 not found: ID does not exist" Oct 08 00:35:26 crc kubenswrapper[4967]: I1008 00:35:26.713477 4967 scope.go:117] "RemoveContainer" containerID="21d17dc5b2a6acf21805f7da541ef94fb231de8098d35b6e2f1304f68d614631" Oct 08 00:35:26 crc kubenswrapper[4967]: E1008 00:35:26.714615 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"21d17dc5b2a6acf21805f7da541ef94fb231de8098d35b6e2f1304f68d614631\": container with ID starting with 21d17dc5b2a6acf21805f7da541ef94fb231de8098d35b6e2f1304f68d614631 not found: ID does not exist" containerID="21d17dc5b2a6acf21805f7da541ef94fb231de8098d35b6e2f1304f68d614631" Oct 08 00:35:26 crc kubenswrapper[4967]: I1008 00:35:26.714656 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21d17dc5b2a6acf21805f7da541ef94fb231de8098d35b6e2f1304f68d614631"} err="failed to get container status \"21d17dc5b2a6acf21805f7da541ef94fb231de8098d35b6e2f1304f68d614631\": rpc error: code = NotFound desc = could not find container \"21d17dc5b2a6acf21805f7da541ef94fb231de8098d35b6e2f1304f68d614631\": container with ID starting with 21d17dc5b2a6acf21805f7da541ef94fb231de8098d35b6e2f1304f68d614631 not found: ID does not exist" Oct 08 00:35:26 crc kubenswrapper[4967]: I1008 00:35:26.775678 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zwkdr\" (UniqueName: \"kubernetes.io/projected/2b4b9b9e-7a8f-43ef-a386-6afc3fe66b0c-kube-api-access-zwkdr\") pod \"2b4b9b9e-7a8f-43ef-a386-6afc3fe66b0c\" (UID: \"2b4b9b9e-7a8f-43ef-a386-6afc3fe66b0c\") " Oct 08 00:35:26 crc kubenswrapper[4967]: I1008 00:35:26.775987 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2b4b9b9e-7a8f-43ef-a386-6afc3fe66b0c-must-gather-output\") pod \"2b4b9b9e-7a8f-43ef-a386-6afc3fe66b0c\" (UID: \"2b4b9b9e-7a8f-43ef-a386-6afc3fe66b0c\") " Oct 08 00:35:26 crc kubenswrapper[4967]: I1008 00:35:26.782318 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b4b9b9e-7a8f-43ef-a386-6afc3fe66b0c-kube-api-access-zwkdr" (OuterVolumeSpecName: "kube-api-access-zwkdr") pod "2b4b9b9e-7a8f-43ef-a386-6afc3fe66b0c" (UID: "2b4b9b9e-7a8f-43ef-a386-6afc3fe66b0c"). InnerVolumeSpecName "kube-api-access-zwkdr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 00:35:26 crc kubenswrapper[4967]: I1008 00:35:26.832054 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b4b9b9e-7a8f-43ef-a386-6afc3fe66b0c-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "2b4b9b9e-7a8f-43ef-a386-6afc3fe66b0c" (UID: "2b4b9b9e-7a8f-43ef-a386-6afc3fe66b0c"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 00:35:26 crc kubenswrapper[4967]: I1008 00:35:26.879890 4967 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2b4b9b9e-7a8f-43ef-a386-6afc3fe66b0c-must-gather-output\") on node \"crc\" DevicePath \"\"" Oct 08 00:35:26 crc kubenswrapper[4967]: I1008 00:35:26.879922 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zwkdr\" (UniqueName: \"kubernetes.io/projected/2b4b9b9e-7a8f-43ef-a386-6afc3fe66b0c-kube-api-access-zwkdr\") on node \"crc\" DevicePath \"\"" Oct 08 00:35:28 crc kubenswrapper[4967]: I1008 00:35:28.137629 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b4b9b9e-7a8f-43ef-a386-6afc3fe66b0c" path="/var/lib/kubelet/pods/2b4b9b9e-7a8f-43ef-a386-6afc3fe66b0c/volumes" Oct 08 00:35:35 crc kubenswrapper[4967]: I1008 00:35:35.118184 4967 scope.go:117] "RemoveContainer" containerID="dd4b0abe7436dc2e0c64a27b31f4a959aef1a772ba57812e2a8fbddc81d33d75" Oct 08 00:35:35 crc kubenswrapper[4967]: E1008 00:35:35.119121 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xvdtp_openshift-machine-config-operator(9461162a-c27d-465e-9ca7-2d3936877cd8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" podUID="9461162a-c27d-465e-9ca7-2d3936877cd8" Oct 08 00:35:50 crc kubenswrapper[4967]: I1008 00:35:50.118592 4967 scope.go:117] "RemoveContainer" containerID="dd4b0abe7436dc2e0c64a27b31f4a959aef1a772ba57812e2a8fbddc81d33d75" Oct 08 00:35:50 crc kubenswrapper[4967]: E1008 00:35:50.120114 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xvdtp_openshift-machine-config-operator(9461162a-c27d-465e-9ca7-2d3936877cd8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" podUID="9461162a-c27d-465e-9ca7-2d3936877cd8" Oct 08 00:36:05 crc kubenswrapper[4967]: I1008 00:36:05.118121 4967 scope.go:117] "RemoveContainer" containerID="dd4b0abe7436dc2e0c64a27b31f4a959aef1a772ba57812e2a8fbddc81d33d75" Oct 08 00:36:05 crc kubenswrapper[4967]: E1008 00:36:05.119037 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xvdtp_openshift-machine-config-operator(9461162a-c27d-465e-9ca7-2d3936877cd8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" podUID="9461162a-c27d-465e-9ca7-2d3936877cd8" Oct 08 00:36:19 crc kubenswrapper[4967]: I1008 00:36:19.117580 4967 scope.go:117] "RemoveContainer" containerID="dd4b0abe7436dc2e0c64a27b31f4a959aef1a772ba57812e2a8fbddc81d33d75" Oct 08 00:36:19 crc kubenswrapper[4967]: E1008 00:36:19.118512 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xvdtp_openshift-machine-config-operator(9461162a-c27d-465e-9ca7-2d3936877cd8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" podUID="9461162a-c27d-465e-9ca7-2d3936877cd8" Oct 08 00:36:34 crc kubenswrapper[4967]: I1008 00:36:34.118329 4967 scope.go:117] "RemoveContainer" containerID="dd4b0abe7436dc2e0c64a27b31f4a959aef1a772ba57812e2a8fbddc81d33d75" Oct 08 00:36:34 crc kubenswrapper[4967]: E1008 00:36:34.119541 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xvdtp_openshift-machine-config-operator(9461162a-c27d-465e-9ca7-2d3936877cd8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" podUID="9461162a-c27d-465e-9ca7-2d3936877cd8" Oct 08 00:36:47 crc kubenswrapper[4967]: I1008 00:36:47.118266 4967 scope.go:117] "RemoveContainer" containerID="dd4b0abe7436dc2e0c64a27b31f4a959aef1a772ba57812e2a8fbddc81d33d75" Oct 08 00:36:47 crc kubenswrapper[4967]: E1008 00:36:47.119310 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xvdtp_openshift-machine-config-operator(9461162a-c27d-465e-9ca7-2d3936877cd8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" podUID="9461162a-c27d-465e-9ca7-2d3936877cd8" Oct 08 00:36:59 crc kubenswrapper[4967]: I1008 00:36:59.118059 4967 scope.go:117] "RemoveContainer" containerID="dd4b0abe7436dc2e0c64a27b31f4a959aef1a772ba57812e2a8fbddc81d33d75" Oct 08 00:36:59 crc kubenswrapper[4967]: E1008 00:36:59.119263 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xvdtp_openshift-machine-config-operator(9461162a-c27d-465e-9ca7-2d3936877cd8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" podUID="9461162a-c27d-465e-9ca7-2d3936877cd8" Oct 08 00:37:13 crc kubenswrapper[4967]: I1008 00:37:13.118164 4967 scope.go:117] "RemoveContainer" containerID="dd4b0abe7436dc2e0c64a27b31f4a959aef1a772ba57812e2a8fbddc81d33d75" Oct 08 00:37:13 crc kubenswrapper[4967]: E1008 00:37:13.119294 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xvdtp_openshift-machine-config-operator(9461162a-c27d-465e-9ca7-2d3936877cd8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" podUID="9461162a-c27d-465e-9ca7-2d3936877cd8" Oct 08 00:37:27 crc kubenswrapper[4967]: I1008 00:37:27.118413 4967 scope.go:117] "RemoveContainer" containerID="dd4b0abe7436dc2e0c64a27b31f4a959aef1a772ba57812e2a8fbddc81d33d75" Oct 08 00:37:27 crc kubenswrapper[4967]: E1008 00:37:27.119339 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xvdtp_openshift-machine-config-operator(9461162a-c27d-465e-9ca7-2d3936877cd8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" podUID="9461162a-c27d-465e-9ca7-2d3936877cd8" Oct 08 00:37:41 crc kubenswrapper[4967]: I1008 00:37:41.121260 4967 scope.go:117] "RemoveContainer" containerID="dd4b0abe7436dc2e0c64a27b31f4a959aef1a772ba57812e2a8fbddc81d33d75" Oct 08 00:37:41 crc kubenswrapper[4967]: E1008 00:37:41.122423 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-xvdtp_openshift-machine-config-operator(9461162a-c27d-465e-9ca7-2d3936877cd8)\"" pod="openshift-machine-config-operator/machine-config-daemon-xvdtp" podUID="9461162a-c27d-465e-9ca7-2d3936877cd8"